var/home/core/zuul-output/0000755000175000017500000000000015153510017014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015153516176015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000176554615153516114020300 0ustar corecoreLikubelet.log_o[;r)Br'o -n(!9%CMc;b[>Ǧ(\XGf͙ *|Fw6l}Wߟn8כțx{wN_ Ç/ixK>|/w1OolW_~~yiw|V/']n{?|v^mRl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~J7FP7M$7iXύ^$26lDt#3{f!f6;R.!$5 J:1*S%V!_F([FbDY娍ԹiY03`Cfw":ɴ@=zN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+׼ s'QWs]8MKf, # qe䧤ꇾ3,!N{\00{B"唄(".V.U) f*g,Z0>?<;^N^iD[NrmN@ Ң`?Tã 5g=XzߛoE䭸[ki|X&po{Wl9HGAr Mme)M,!])V_帛AB}nyи0stĈCo.:wAZ{sy:7qsWctx{ul-+ZYsI{o.Ra97XcђQ0FK@aEDO2es ׇ# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[< ~6m)0}(*T7Siv'=k 9ԻreFj?wQ+KmrI,W i̸.#v0nFNV-y(&e'd,LFlPh ۬rW-V5e1߯F1>9r;:J_T{*T\hVQxi0LZD T{ oWHWc&)_`i=į`Pír JwJ`}OPSSii4wT  (Hnm//sE炱}r4(9qfhs8u'8KwI~3v4&8[q_5.)Q VE JN`:a!KM/ 9G uIo1]ߔr TGGJ\ C.eTEZ{(¹:-“lՐ0A_F叻l_}zD}c6  %T%St{+J{=v-}҅"o ']쌕|tOX8nJ*A*%J[T2pI1Je;s__[,Ҩ38_ь ͰM0ImIMiVJ4&jN'Bx)0v߁R[F)RH?uvͪ_5l *7h?aF_VxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mp+ھ*V I{a 0Ҟ>͏ ,ȣw§`Ee$Ə{(he NSfX1982THwnUC9fDx5X@O5ޔL3VQ 7,oT5/tMJ%\t=[ٹ:11:2`c J1bV_gɊ:+^V,~0{gj"A, rXr*0ngY.] <ʜ6 ;,9VPAHuŠ7կhw=m{> *nacԇ&~hb[n㉫k:%݌6od FT'BTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8fq CXReQP2$TbgK !)CG?5rk{-cS`y_B}V v,{*1ߎ% qƦat:=G=vNv dߋ{Ny[$ {ɴ6hOI']TF>0st\t@HTu( v e`H*{Ögڌ:8cN|U1,-N9 dI [@3YN%:ò6PT:”QVay V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bd+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:16.Fٱq1M k'JE%"2.*""]8yܑ4> >X1 smD) ̙TީXfnOFg㧤[Lo)[fLPBRB+x7{{? ףro_nն-2n6 Ym^]IL'M+;U t>x]U5g B(, qA9r;$IN&CM(F+ hGI~Q<웰[, qnriY]3_P${,<\V}7T g6Zapto}PhS/b&X0$Ba{a`W%ATevoYFF"4En.O8ϵq\FOXƀf qbTLhlw?8p@{]oOtsϑ`94t1!F PI;i`ޮMLX7sTGP7^s08p15w q o(uLYQB_dWoc0a#K1P,8]P)\wEZ(VҠQBT^e^0F;)CtT+{`Bh"% !.bBQPnT4ƈRa[F=3}+BVE~8R{3,>0|:,5j358W]>!Q1"6oT[ҟ^T;725Xa+wqlR)<#!9!籈K*:!@NI^S"H=ofLx _lp ꖚӜ3C 4dM @x>ۙZh _uoֺip&1ڙʪ4\RF_04H8@>fXmpLJ5jRS}_D U4x[c) ,`̔Dvckk5Ťã0le۞]o~oW(91ݧ$uxp/Cq6Un9%ZxðvGL qG $ X:w06 E=oWlzN7st˪C:?*|kިfc]| &ب^[%F%LI<0(씖;4A\`TQ.b0NH;ݹ/n -3!: _Jq#Bh^4p|-G7|ڸ=Bx)kre_f |Nm8p5H!jR@Aiߒ߈ۥLFTk"5l9O'ϓl5x|_®&&n]#r̥jOڧK)lsXg\{Md-% >~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==q?{WƱPz;| \;_D[T/BI GH8@"t*"9z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ35dEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1YןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄT/K*!<+"eK5c&`X:#;@B@[(K44sBFu M.MNWLlY]K᜴=/ VމYlϿ4i36$>m|_>9|dUA"{!$jKx E$K3hN(tÊ-#v#O N, 9g80Ǭ&VdӞ5W1!1KYd`,-*&>F~⯰&jb.~cNk BL_OG]Bv.A|'qT(Ol.' 4IE|@Iі)<-p JkQm1 `qacܗVc?)cl*&<}P媠E{-sVU>߇GUt\+n3X]Byoz)li$2cPs6D>TE-n# rve{椱I |p)U݋7yJw&PzDgi xs  xh\L r Ѥo Zt(I >|$>tnMdэoe:9[_v~\:P ؇'k01Q1jlX)/ΏL+NhBUx~Ga>Z"Q_wjTLRˀtL L+BT҂ll魳cf[L̎`;rK+S- (J[(6 b F? ZvƂcW+dˍ-m𢛲@ms~}3ɱ© R$ T5%:zZ甎܋)`ŰJ38!;NfHohVbK :S50exU}W`upHЍE_fNTU*q%bq@/5q0);F74~'*z[\M-~#aSmMÉB2Nnʇ)bAg`u2t"8U [tJYSk, "vu\h1Yhl~[mhm+F(g 6+YtHgd/}7m]Q!Mę5bR!JbV>&w6οH+NL$]p>8UU>Ѫg39Yg>OF9V?SAT~:gGt $*}aQ.Zi~%K\rfm$%ɪq(%W>*Hg>KStE)KS1z2"h%^NEN?  hxnd/)O{,:خcX1nIaJ/t4J\bƀWc-d4M^d/ ʂK0`v%"s#PCoT/*,:[4b=]N&, ,B82^WK9EHLPm))2.9ȱ  QAcBC-|$M\^B!`}M^t+C~Lb }D>{N{Vt)tpDN,FCz~$)*417l;V iэ(_,j]$9O+/Sh]ice wy\Mڗ$,DJ|lj*à␻,?XAe0bX@ h0[}BU0v']#Vo !ې: Z%ƶ(fl>'"Bg< 0^_d0Y@2!ӸfZ{Ibi/^cygwדzY'Ź$:fr;)ٔf ՠ3Kcxwg*EQU{$Sڸ3x~ 5clgSAW"X Pҿ.ظwyV}̒KX9U1>V..W%GX +Uvzg=npu{do#Vb4ra\sNC/T"*!k愨}plm@+@gSUX覽t01:)6kSL9Ug6rEr(3{ xRP8_S( $?uk| ]bP\vۗ晋cgLz2r~MMp!~~h?ljUc>rw}xxݸǻ*Wu{}M?\GSߋ2ꮺ5w"7U0)lۨB0ח*zW߬V}Z۫ܨJ<]B=\>V7¯8nq~q?A-?T_qOq?5-3 |q|w.dަ'/Y?> (<2y. ">8YAC| w&5fɹ(ȊVã50z)la.~LlQx[b&Pĥx BjIKn"@+z'}ũrDks^F\`%Di5~cZ*sXLqQ$q6v+jRcepO}[ s\VF5vROq%mX-RÈlб 6jf/AfN vRPػ.6<'"6dv .z{I>|&ׇ4Ăw4 [P{]"}r1殲)ߚA 2J1SGpw>ٕQѱ vb;pV ^WO+į1tq61W vzZ U'=҅}rZ:T#\_:ď);KX!LHuQ (6c94Ce|u$4a?"1] `Wa+m𢛲`Rs _I@U8jxɕͽf3[Pg%,IR Ř`QbmүcH&CLlvLҼé1ivGgJ+u7Τ!ljK1SpHR>:YF2cU(77eGG\ m#Tvmە8[,)4\\=V~?C~>_) cxF;;Ds'n [&8NJP5H2Զj{RC>he:ա+e/.I0\lWoӊĭYcxN^SPiMrFI_"*l§,̀+ å} .[c&SX( ( =X?D5ۙ@m cEpR?H0F>v6A*:W?*nzfw*B#d[se$U>tLNÔ+XX߇`cu0:U[tp^}{>H4z 4 (DtH-ʐ?sk7iIbΏ%T}v}e{aBs˞L=ilNeb]nltwfCEI"*S k`u ygz[~S [j3+sE.,uDΡ1R:Vݐ/CBc˾] shGՙf 2+);W{@dlG)%عF&4D&u.Im9c$A$Dfj-ء^6&#OȯTgرBӆI t[ 5)l>MR2ǂv JpU1cJpրj&*ߗEЍ0U#X) bpNVYSD1౱UR}UR,:lơ2<8"˓MlA2 KvP8 I7D Oj>;V|a|`U>D*KS;|:xI/ió21׭ȦS!e^t+28b$d:z4 .}gRcƈ^ʮC^0l[hl"য*6 ny!HQ=GOf"8vAq&*țTOWse~ (5TX%/8vS:w}[ą qf2Lυi lm/+QD4t.P*2V J`\g2%tJ4vX[7g"z{1|\*& >Vv:V^S7{{u%[^g=pn]Y#&ߓTί_z7e&ӃCx;xLh+NOEp";SB/eWٹ`64F 2AhF{Ɩ;>87DǍ-~e;\26Lة:*mUAN=VޮL> jwB}ѹ .MVfz0Ïd0l?7- }|>TT%9d-9UK=&l&~g&i"L{vrQۻou}q}hn+.{pWEqws]]|/ǫ\}/J.MLmc ԗWrU}/Ǜ+sYn[ﯾeywyY]]¨Kpx c./mo;ߟRy*4݀wm&8֨Or4 &+Bs=8'kP 3 |}44S8UXi;f;VE7e4AdX-fS烠1Uܦ$lznlq"җ^s RTn|RKm;ԻZ3)`S!9| ?}m*2@"G{yZ${˪A6yq>Elq*E< NX9@: Ih~|Y4sopp|v1f2춓t$ėƩb{W6_K(>XǙ/N7`Fݔz+zL'x8޵ˌ.?B^s9q_rp^tJʰk))[-tOlɩVs׶.C0׶iDXJًH,GkUŖ'׵*4Xx7BfA䇯kUc;׵HM#l#nF95z1chҚir4Yqd>Kъg,yr~V2,Md\aZ |D$ k2xoMV< ZI7q_uM\2ˏ_Yg0-d;Q$n/e9d0^>Na8{M: s]ɭ 6mpCst?025e0("7|$;ZRku k\X_2]|c|6\i_\X }?Y(kfN50A` &28ߴk|ۦkZYO&OoXW[kT{Koߊ9kx%߰ר ۅX1d$pɏ .n;*F27V?q5u;%2gnlF j.}Er6jǒf.dUՋD/U];jpOa!ZR5Z+ʼk*̈msYۈ 3-# M˲0b?0|kԋ͓#2̮ 4~<ȊUwGHB*ѴҴWB/r Bd. X3BV,*'"o` OMꅘd֤*D><a~L'BLu]TogdQ5YuY]s]jW9a59xJ-ËxT-q&DK.X̒*_oE'B5]=ۯvu`(sq\eM oy7&&Q-Npp X` Q,a8:Нvvu.Y w9;;TAC}).ʛ2 k _bJn59s7 Y۝#g2-^A4VxG49 ;#>*//K6ߓQu^14}g![vVݿoXr^nsqrV}GsvWM%OI ݋ FޭfZ/VI|swC8RYpt{T#p纍~(8XQ-rKjSSL^ 3Ar\E˯ݕqT%x $@E8LzuAG_O4(I Ǵ"Jˈ| |[arNU ;sI'DC }yhsTapйcǠAҟX*0,iN{X昅`3FyIrOq1oMRNY&UI oǸu9o|w]/j`N-o1 ;& ,GK6!wp@Ckw.Ѕh|/{k+8d<"ej QfR2T&-z ug=Vn\Hh.KORvs쨩*4(? 0yhH&I0NgTG R f`a xݙ҆Y:A5s$5A09(*,TpRAL P@qAw>?a𨑁A6Ʋ(|8>A H@N!eU6EΓNdKW LخzVI؊P7c"p&~<0t"7LF+ŀN/E]p症1K*" lkwZ;ThH6!7~;@%W5gp$ZV8dVh> 'xN&y[}x0+-i.A;[ `?k-γglE{tuz`2(d lj)+foCk )A1kFҍ^ @%X 2-!OjKڒTSߖzudJ]nIHpS .YV~Ȕe[5[3fe[%cײր0 okRy9&ZZ,/Ś+ÀH x_ʱ ,t?`DWq/_ "TܣsZetŜum{HĭJ:BW7Yt'FBJS{3@͡Q,D*,Mo/hlY؆"FV'U@U?IZFħgUs+T哑\u. \Z΋<7?Qvg3HXY{$_E9Ò45K3wZjWqY9h{fB3g͵lGf8- [kfhl>}ByVLIDhޓw_bp6"4'wE޾ rPfw6F0|,OM: M;؃suj BP=C8}DPS kOs#zz&e6@ ]s li14jcY{4r{5>M[C@K):F0ORoCLfB0/73M؏ 80(! *bEP&&fsHrSսk߫6`LFVݸ/Kj54̾F ~L(ޞ1A(L,HpCOӊ%uO^!/3v3DrgXu/Kwd3`1RgχFBtyndwn(K6I ^;G({s/ 3BޝCegqiPvqΏr)IYMGê.x~kcsaXqH"Xg0dFگyEuEtRHF& `'=#f(i9OUi4ˍyRдqnV 3sw/ |qRONڛi2l6#%;LH " tI|vv%\BP(ѧ&D~R!뫽qs(@)?O_O_WO!|}~?L1fCi>TB9&=~0IgRKNal+ua?eX_zPI]#e{ǯKUVeߗiI7'Ql.x5tQJE-&idU:\P;g?7U6x-iȺ]vM[ok:0D97XE^}}rd|{rzwy|r/{9X^Zp8DۭKؤ;h Hk%@1s3ng^Ͻ{ u@mi~7|c3Rf$@1=hI8yf@mI@Y;܍ 8ے |whmI`sns}7I֬O.`1PyN2`L&#~ʾmF6MZ5!L- -%0q N  =[ͦ뺴zmǍ-(~=T3HRQ6:KOXM:pT.283Wjσ6^q-.7IkħsJbQrZGj"snPbaִ4q Z`k| F0ႵA8N[0Xs`Yk֥}uu˳_Wpad"hK+7%mS,֎(֎w 9g!.OP0JhxvnӖ Ӷ G<)muq䡁 j 3)RjA#YIj/i9q+胸dܡ+X y -s("iQY|@)I?gbECܻ0|g,7_2<@C߆ج\b{B[x݉EM Cخ/fPʂIJ5Ӱ$6c);#/D#JM `[:o {-}I0aާL "kmh&L~v' vTI0{ )LD& vj\`Re7%+:$SpkT RǸ&&x]qὔ+YZ,GB:i:VwR fGF9$"f8Qfi 4YE,t.d4@2<'о .2OW1s?&]W@[QxȂsU.+rS "IO,F\䰢~nIz',NUס({ w? I]zگ/?X!ѐL qDڬ+.T"qO.@-* 0?h^\O.%rs'aپv OLc XpɄ̄$*ӧ4Cw.SUURU>t jg/AGO3?~:}';E"~<wDn8;)ڪ@Y؄y'Nu`1Jm7}i`ia\@3IfŜ΋;MHwu%QnEB~; P:NX?EhQ6 6t46tHȜP窚\܌ 7ku+װt-ׇĸۈ=X*uūFݭgM{[RƟULMi"TG`#Y^N# }1wf8.B ڠ:pN.8{`#>>IF'z ~qPN1+iy'+w:dx,q|ϥ0#]<QAG wU8L^9OW\~ &/S I|o_ ,bT4-bP/(Fi;~4&Wr!%0iD xI'ݸxl.—Pg}ډAq+:Q?S qj R23T+u(jsI߁1 2t-(,?Ge}`O]x=,d|9+\Vt.c KXEЀ0nGn>zO .1oѕX: <ŖpMz$XV9GWkD9[}ڲ>Hp!Җ-ҧɄ\l Z,F}Y#}:&rm`x~Sh$ TW6QlE6Mf`| a]"&\褟q' !t`+P\36.uU<&PVa/d.uua;pvC׺|sU?^ ¥zn[rAt˵ڱUCn6\[A՞U}x(vG)w&K Q o =:V]٩=>$P89HLd-틍Vj:yw dYai.:X261V$U9ޏc,gÆTTn_~_"}~J ;Jw'F(ݓP +CJwSǠG;tq]؝Fv=tl]F-Sv N+;T'lBeBuv'9PgOBu uw ݝP0B= uw [&ہPowB$ہPPB #ߓPB2|B{w T,*v TN8P'bBeB v'48`OB .fVK ?1^Pn!mFlzϦ$TR*G.&Hz'/2-J)kJKsH<œB){]vx׳kY׾1\(۲. s4%Hʔg)nkq6:׽# .2ߜQQ$L)F`\[6]vtLI!_3IfnprdMo7ʎB95V ;EM޲vRU,TP;O'ʩK 1g ze:(fZTY"}jZ\o%uD^X\KQ`(hLHʇq,uH5 7pe4y)Ia'sF8TœgH_7X#n-.2Д|Qfjƥ00I}&v!%R) Qf $9L.%2[L^kW;Y' uh6K 6M6MCB~[,Y&tT˽^a^TLt0kr(Ib[ 0?@&_~PvHo!?h tJvEg-|\ ZXK8p^#**J`wm)<@<M={ʼ4V<)QOրPhޝJr÷dSK7C,r]0 T/P::J@UN=@$<]RԪ&XgBx*@c@t'3VYk{Q5̀iBY?"fDwӎԼ iRƐƃ z6LbmP2.338J}T$fi 668`*@yM y :+M i3!EͰwé6)x%:l8nZtT]A/sBk N{⏧ MǠ .0Nd.^u^~|"{_ nWc+9 9P(UF84a/i*ueiMag t4T^--?٤?+JN6*\,X䡰2Ndx5g<̫PzGydc9Ch0 BlwHz#eq34ڦcwzɏGo~ܦ1l|oY5Ky5S)uCGheu_eߔյH9a)nJ:3I^ȕ>E\d޳gk54 \uޚs<5Ydu3l 5C244>Z-ڣyT ÔUjb2zQLUjdȬ[ycO`{Me5W2oBht @UOIJ2mЍny4a>>DCm-s-W8jϪ@T8h0EOaNWyVDdGʑlAVgc "3;Q׍1|7&mӎ`"UUDkmH%ȳzyX b/dw}Q$CҒH8qs:bYtRXuo7z4m? qLŪ|YWb̏~vӉ=ɜOom?[/)> 5œ\]C/n|f+:o6=,V'F!'aK˛stͿ#ډNᵟj϶,y8`iZVQYW}qzAAorO-A*=at >Y N!깣?^ruLt,-e!T g)jcp%1쩦ǾM3yMGGCVZ %6rM䖳M_5#HYUDZ.ddVpo5D_ 2)[i\gكۄ55vU/$ЍN!c| |>,b̃ R N[wDוQ3-^_߬/L2szUNlZ"$}'Wb$OvN ~/aJX71`*G1܎5rT݅F.f</e&j{)Z? FLo $`w//tW_k6TNAO*[K+g'-*L:c$]$K>z %\S;,AX`8kV&`\CcòJ%[-mI|>88FX{B<ӧ42e-ȒF[}>& xV]UX7iwDvuӆG76ѥDɁhЂ}w<ҎD>|Fes\ Qو[n)MU^;b[yX$vk|o,yxIc-G"$L9˯r:.\i9쵱`kopQtn+o1R@hsY2&69lNYm0Pwgs!qBmb[t(N9A qK\a וdWNНj`Fq˫"c=[뀪M<ARkS͡N>09Է8|n?;m?L\JIC|?MuWi`"ͮ\9,dZ,>PI0t;Y$8Xinp:Og"\t?CTd o(svNʫ⦘м)hNJHF1~pYdZ5k:s])gSJ598rG [F !M5s)XOQUd4H-r줺*&I巐ӕrWۚv r[ fZtlEh}b02Gms}3v,0-e ʅԤ7`Z\Jp:BDiF ~e>Jܓ$ m@B0=O&@]$[~`f W!sV /drK:įӣpprҤU SQ5ONpr_KH!nuHIK %F |oA<VMUW;bc'5%+B+Mc|$b.PY*1al[Ig {F=<7BWw=osWM0.3Fsbg O3KYQlU%l!r.Ip6x2{:@Ip*ݾ=[%mk $Pﺒ|Wri,B^>U<ZL-Ie!3F҅>6hUYUX3[#|wAfa0؎ey$ _;90eʼ2Yoh{4cTd/tnXuf+rt#|% syLЌ 1gR+!eʂ' Fҥ?W$8n༅ CY~]jpǍK ,{ٳ PR(ĸ㌑jb{EeBVԄ;V"BthVn&-Ұ $w9x5z.%שŒzn;UΔi 1CdC.Jeܒ8zn<-ԥy=8S;ݗ]/Y\.OŎ9ù,{a܉W=/w~w$qכn3KbV5o= QjVKz\$k6Z#@K_7: OwO ?~xE:.ؗi݇oDF+SȠ^shh.^hV| NWn+9r!}SpLüsK(&lSˆaYq}V 1x,T)o.XX[Vb+.WBsY@ C/\}-QpTcq g[e.0qƧ|!#B^IpL8Mu2!T}.\ ʪ-u.8x}9_-$WL,4ӡ<9N!%98$hcvNXe2"$]vܖ{Cc*OC4 pd$49q`Ll*6E]6wKOM6L/&f O*-_,ctaHp8 0~ nXCS۬LY*k)VR:'B.K5bO N ސ67Lb!i5{h*pUN1-)/ZдKbRC'qkpw$8&2?'*OII}ULtQLc\,cZVj[}G oOj-$8j۬7uuw`6v%&qp3䬉(jb28z|8㗹NTa085$d=8~)h2t%g))p4hXԘ/E6 rCMq~BwFpOꇶX-vw!zÍϿJ8 v޵qc2vQۼ_ Ed f~X,yuI]EN$yt]v ys.<$c axܷ~:R7A3iTcUUQ( f:5 \ RLI2N+ Gq~T#+0bbw8?5X"sߌ^'67=_6Vkg!Ge1 yrзz7?;N979wq?26Ffguk.Ua7n GGWIQ#eRf*Ku[ %YHb<0tdGD씰mI$I$h6Y,s-1`ˢ>wt(4K$(]*[0]H F z~ug&=s{}{N ^GYz[2o~惟ⓟ~[YL/?IHgɃo<±a( 6`?WU9w~te^$O+E&>㟪10i޻N'''1$=7.`7&'?D#ؐll/ zyRLuA1FXcpj4Z!: @4+'v]r* ~:aߦN]IXvrJ+ bVjr4orAr~ګڻ&St䶬 X9e6zY38*-½n{䗛rԈI s/E\*N<B x54jN9,FuoW+i@Mk.U3hJCX8\|&m; ]dzv7ۏ~@;ا]*۵KmǕJI{2#O4䭢DZ= ۓ^T] #ٚ!SIf̪,S񔭙@| 'J%$8&HNbeqܥS;EqUgAZAQ{*u Bxa7Dsź{$8gˏZ΅`<.%׋-O˃/o"^_tOQܱwx2[Q9O}#.?Gq&%6 ̑N ^R,X1W)s ]D&MZ6u-w;ƽ.F8U0=8w"%ڡt7 >3Yi̋k,y)sj8Y_WwY3RrA>o7֒ ϬXJ P궔Nq=0C6(xD?#6[>-6l:je/WOa7P8uvQ_ ַIhn3YTWi/V_`o1+%8FYBe&02OJCd"j!B)b6I9 w 7;a 0VmexKYIJ:$/,JC$® 8%J4%8X٥d"5ij?iFZӘ4v@\ƻCK[HJ3`Ad`K%.RkCC F"ؖO::|l|b擈N1Du@\Z*q)5[9O(3F +,d  g5 U:<6@WqqsسUSJcJ9|HIi"apW=5C*cR >X#Q90PRM+D$Z B7 E PL)L;`SÃDMlj\Ife`iMV %c( 1Q\%ciϹIȳ(u1:P',–|tE" Y 3TZłz+!O|[0g֍nƃ݄")c\u^guAi>SZMI{EiJmkF^-c#mk{\_%\ ̜ vtMCc_~&-k*1@Fޟf zL&?{8e7kèx1 Q:ÕsʔP2 udTݝ)0y4@<2J%FcawxyDrXYjiց H )JPaRLiBh8,7RC6F~f:,?)9-1tc啵1`JҔvb.tp&`D%j21ô,*r*xJLaR0uQNlв!|t>,t0UP(!S0- T^;O(k݆1e`NP 6?be<:B{v(E# -q#L0Oe1f@Zx.x,`-AT!x(RIݑ$k&r-`ֱNuȍeTwǯ`7vm &ezt%YnK8% glđ K  =kńtƒCՄštau m&0ca*0BNiwHC뢴 =kQs|YVDLh" U]JC)(]1 :> MR:aEԌ'BQOYo hR`;L0pR2:^qNVI CB@R>0)w%mvcExއ\Hx$!IE*52KNh;ehQRY8%1Iz)5|T֬-voOύdN-uF 0)LԄތEwj60) Qwg60efC} \E;4ݚ ZuJ6hxWPzV ƻBJg)Z 2diFTnu&Tg:cQFI00<'MX\8RiRٔhi+ ?Jl# $yWiq.-A&`~߳#%"i<i6*U6&4I&qDMLR* *$jJ*R1ny2(QBxr*% ̫TPM #G qeB7#"P0;uz77pR8#MPh  Ef+A|zՈ"kա1Y.5]YZǬ34.W,&涅į5f&Q&W8sb3%D*Ƒj%RB5#"b0E` +›18 bJT+=ǡUSAwB6mQ#bBY{YhHʸ2Bwg*@2Wv\Eȹd4'Rh!TtҶQc.]v B{BQ*T' ۅ8%fЛ(h#:5LP\ͺ34S*&;4+ViD>,UxmyDOMrH*f0Swte2k`Hk\:7lvi6XA3Sq:4}be=#p$g*70(:_Y82Ua)+CUؕV>4f\/AMl1_Gbͫ9Wsݕa!x1GȊ旮ػBT#x[vj)kZU8t /qr]~l4sս ^U)nk3Z\^-nw12uD4z\zaC];̲\Itd޿* }8^;z]d$Y.nnWlpiz hMCKArTE=?xu㌶kw[K +Pa}ͯ&r;\FRB\G%a) oLex>\FžPϜ`*%.[ ?2C6:5 Sޖ.rY|f3 X2{MԿVvMQ8YTu!熒﻽V95\fNm(Z9SoMr uݿ}7sz =M-Rq Ҳ啲y?u7VlkAZw  9Ѐ7fU&Va-9_r`ˁ[g?Fc;pˏwذe#xmAV+ܤ"bB?]UdR6!וdٵ]FG;ml`Z!»H8&p,]q&7O`*c\jm,on‘-ko>&髬Vj btbWЋ .f`_BY FJ^Lۋz tM7U. Yț (!ϟxp8}><(}_m 1L,iv1 aί9EӋ̜Y삄O?OY0ūy|tx2/1;Sd~ۆ,xoOV{k{*JD7I-]W@qwAYrTM_?Ee{a/`mPn0,ɧ!UY lt:P._4koQI `GEyTZ@mQk)Ɍ0Ťoj;s2 %U{* Uv_jn.kAEA6ڎ1owTg)=q?? ?}rƌ&2*Ds9B!Qw7 ~}b?߫ϣV|c?^f9XI韠GΑ:`;RG×+`ts:: :}:*4RW/?X*2*R<Ũ0; 5tr9qtb!1.F#G)qcC^LrCxhrX=yc`Rfp"Qt1->{]N BX- +)x[SK "lMͷm ?ac﹄]=N?g8g5"µ;J[ pNj1?i^":ZL#V|1@*.caV0OMcwHi˅:ڃ 0.g)!CX32R{c~uޔ WpZhj渀@&wEK+-cEhd.qנa6Ж>pek|t'6A ;j7~Խ&h'zb7"WNQr=\L?driӓ%,OEx:!<3WC9ia85,K&@˪9Yզf G ^0ҫ)Vj3\0Oi:\95zWKBZnږ6p2M0u=:y,;.ݾ AbswvCn ڳ1x [~KV }~/ #  zG؇"r|Už=tޢg"?yS]s~PmE-զ#tԠm-49.AëkQOi83eS9Ft8Ӽ l]jФ] jFDx LNhx`Qn8ŀZj%?;,~>T`+|NpbNaX^28pYaE3'/J>+Gpg y>dj5ݜV/]RINܣ lNvbg XXGgza.3JIEdzSyS}Ɇ}eo0Sv%wSRL 7S @vppY rgjJuo$b~P+) USi aaR,Ժs!BzW+D,J0Ϧ PPGq@g; Ĭp+v@O#thaDt93 !\aQ0.Hh vّpWƠUj-u+cE+{To+{onjѾ؞q.t_Tt_+'=1WqSg~a+<0Wۉnz~ aZ:?Ώ>]l$_7ZX!B/@pa482p& ;$A9)/=<>ijs(L^#y=K7pR'dTz}UiakO^y #IY_ym+بF=ږC$o@ &%`̄fM/8 ĭꪛdO5xp4/K啯 z t-Ӄò\cTRKL*IRqQ"IpxuȈ+!s+~ UÊ2Zu]m,MPbXYD\Xĝ^\$-_4x?c?ԽoZk1I@ 6)[.RlXoW  #huwʿ۰m[1EQ1_l(c$T-1JΦ踭isu؈|Ui%AD$" !i)b5Qb%ImF]U" 3&'@ 0,; 8+*Β>mkpόn0WgAW{&q̈@MqD^$҈Qm?~qD}"D>)iVif7PoQ0K;d8#]􁮭 Yef(JX!JqJ.#5}偮AR{`rU{|X0TXHu``2( 0W=l=*.7LŁ}J?`Dq,˫۶"kǛk70ȖO7[?WVc\=JQX2~8K#nrTF͌Q+Fr!%,@Ө 6,GdcQ#2XA: `K: 5$t,gf/we}N!Õq@ m06}F/bВZFuo)ZG:e&nhjɃn6Pk. &#:LIhGɮXHLrZc jnB} 61:S5nV}+ٽO0o%/E+ih%]%!kzڏQ Jp"kAeT= nByr s?wJmAޙ'n'Bv \IxUZ50.jwB9 Xb& IYCp>9:'pfSuMxskqr9׌K{MUUQ<25ڸ?Ug A êVs7d"Е!7t`dZ y-3fcBbhEaN_Q0<+ v@5dw y9$5+OVڳȓ(<1՛\X=q0U|j2QMa j)d LNd~`?D \r V0!C! O\Ve\z#ɑ;U>I1/VJ5,*SJbq%eTf1?F5V Nȉ <4%=wB'vi+3K{ׄw-]' `g*~^/09;Gl=/V_wrQ8kr*\t7w7BFh=ZW1% {Ԃm^V碔Fnj?WSǃ=:\n\?;*λdF6jcyA^G5[|܇hˡQsj>LKǃqc_t)\2&d~՚9޾yÖ7%nKݭ%38d0d7`](%$xX6]nV l6ycXٽmvٕw^[q'@;?z&;/F*=G)g&bxy7L6ymcJL,Ǹ! Aφ IX.Xd<ҳo5 ~:>pC]?U_/CXRE7tR}AIv3C+; 1:a*،oVռլ'9w 1Lk\+Y)T}ƠB)ư"j^b~ѫu8"yl j8|P9|p8w/Wm|V;YûNfYB:b- 5&GAdSj}0WRTEVX|!t/U}eCÐj 6(،2chT9ɔ,=0dZЃ BHrc"x0x6"U:(noq{%'[Jg:hA0^R)+^oz%0c*q Gְ@qz{xN/4QtQ`[|ރy=/ƦSKL3j0-NA:qsx=@׋ 2*RH9wZe91ê TZ1-n!#AN2EoKCq{e>.?&g$*/noq{%heD lB*Tk=[Cr{|1zcr ^R׍\Jgm9뫎ET~ZD)},rAa |q{{0no0DGc"E,#Rr\=\78 ܁Tƨf$j[ރt{1zV1jUxw۴/24Kx|ߦ)ϗ@7N7ҬZt!\on_Uל=7{/O?۷G<龯§mesuݟxnTXg S_t];5 hkF_bE.V';\{ձ% mqc]UnM)Ƀ*UY̍G]᪵%cV̭۶QirWλZ":f=JvU: Qo] WyUPczW?Bk<jH>i( 8.eO-mci~hH+MrnՍ:mu9ӂWj +HQ|}L߀C X0m4 p-H'i'2 < w፭ji5qLh5Oz † h!>!^R6m$a~\/!.!v?'z;Pp+RA۴l`l4PCEUK@ܷ~ν+pE{_cȘ8KKޡ!d`4-bEMT/^#ݿS=^{P3 l5]|<hIF8ǔtrh+s#DIH>'V̖>elĒ#f.-H>bLؾUH6P2 flܣ-X%aX و$#pD# 5Cc .V0^\eZlg{HRbT/F=aO&)Q8m$![J,r,@ndV/c0lq$$Lɇ#@lY(u%dFmTBLɇ|( "tvaw lSBq/@D]d g5*!DHҒ.!C⛾*=,_٘>$NMMP.VCmZkD FyPYN=WjFQs1IkF򁘏> \SIؑjS.6 li#Yy!l˶lo[^Xy,"lc=֘#ٰ,gls-#uڅ%;eɴ_ j 뉶1$~+),6K’)Hn Kvp{t׷7?^/߭My3֡(Tuck]-N)uGZ#N>zl-WuQ C@Yv$Mɇr" ^y*Qe E'֮mA+ޤxQ Fy̗xQ{˫K g$vS#+Ѧz/|=_R߱]>hP`G@ELFB&dCXWrR/NPQWhoڽ>jo$l'rg*&ЂMʹ|hC6/ƙl];m^6 ~ ktfIͤlhC&;ڂ>nRзB]wOW7-<3A0FU IF fu] U]E3ǖӈFC2,JTy땿I bslotMgNreo狐z8-B:/ -(H>"ҪnK]4}ťU] ?hk.UAP %FYB ܙo)qI>W6g,9F$[&ړ磻I%Rc"h9>S9>PJYJYeDN:j@.MS6yoKbmNJCbʲ }@0i$8@sp۪ M+;eAV6J:MC54M-K1DhxԊ4p[B{ς  95ט- VFr [<%%/ķ]i\VA0a$p耗2MЖBXgO0==%Ad$vX<޵uOyN|o9O}?B YJ|dN.6-;Z geco(FPؤ`H6 p9 _ի XGm7B =P81#gq}{ Q"Q:UPk  >Y6HZ[ۜ6 @_jۅ՞sghmP KL04f^R|W6sGw P=$@L֔MlhPQ @͈,N4fHĴ|ĖM@Tu[5.f.AG-":C"ĭI#pCP%QsWڇ8]2}yg's$aw3IИ44t(A#DSjhq~EFN68 jإW d-4)VԽwZ@k^Q3CD:Ȇ,[N 4:u%ͳHu+G]!* Wu2i$0MéJ?e򺵥}"Kf.Ђ:)dXR Bf"M=7?$eUʈ$jH>"'%ULօwu}e]ȓ7`K 2i$ d]kfgD˖%@RAZQ%ЦV |٬ 16+((fS dcb%׶ t)E|@ʆMP`h\$엮i<+IʐF."nH aPܡut*Bqvͯd4(XU[/xG)6 (9z=8X U 8VnD%TBw$#ZZHД%!jI<k%D L( $b#rux?A%0ԧcB;$#\籣=1sțM7%۸U=vʊA@{wLI2 S3HGo1` SF 5Krmi,RlaS`$!sB}_<"41!$ S'd#dg93ա"wS\! 4GFF KeH>@%f Uh?Jc+kBp5 N09SW]/T$-4~ JaZS'F€m4us뀞77Ža 90". ,FlxQ KbIKH2c `=m#ApnWgX %%λ!(zH6F}I%2֮K^m_RcxqwI PӸ\w`{U, ,#I\4W *\y+Bgl~YOO/]uq>|n??߾)&ڨړ?_t'TQ-TڭD|Ͽ\GD\~=t]xUp~R/_ie;5gl?{8øf7dn{w3;ۘw4(^Wd[#:<,X,~UEIҭR$)N5Ky{aT@؃.-ߝckW8}_6Plf\׵H YH!\)h0A,噑~&!5ex@r>Rdt'nU8,%xJ{t[1 .2]Ş/v9ҽOw+:+&,kt5Ev'-eР@̓*S^*`|Z= ЇUǷ'̘Τ>d/ :+d%ʦwq\loLJD3YkUuTЗu(NZeJ8_Ćuxft P*`MVůuёHNқ̆`"sAH0iA̷'ꢙ!OJio/Һ캃jd&Sf)u*x)rs <&Jw]w]H6'J(E{bBD+} B6tɻ]n{ ɻ4¹tJa\Fk+DŽ]o%d)lYREύ,c%0;?c;?cB)+*R),BIH(}]~l؟Xu~lv~l.R(`בeF{e1e>بc3W#l|d^^WIԪrt䂮! #KYKيE+i ?LH'v u1qvLIf5QLD:i&&R$skMH \ 괰Qx<8.$\:fW>pƼ)oMCL*Nz7[Gz%Y|K9p.$<= {J$i Ҽ AJ&hm݋䞲n{'=-1j{4r&^%*t 1DK͑vsi7GHyQ,rR0f]sF(!9n#Hd u\(qnlP^޹۹XKQL|V抢[Rxn|wU'T4N? k$ t\??h gy>`#ȵBs' ^{vEC7GY~W}w?>]}x{ i Db,\zFṁiz+mM*{s|LNSQ/|tz sGx__K?Oln~X0}t? PՂ7Z>|}f/Zh% QT[Tӱ0sw6TlbKKGa Mn*^rR$c1kuHTQ ՙ4kJLĸ)1sK \Cb>t#u=J z ˸8ȍn-+_eG>/!xhuuIdҰEy#$niGCw@bbsq:p{&*H'\>, 1fҊpB~xow?fC Zٖ%@{8-ςusAw?JEiH56 V]\>p)M1:[n*W3C=7ood%M-N\e~BCQǙ$dTf(˴2IfMD'̊wJOiG ?9emkmdV gYPC@c) -8!B* LKaXz0\ hM4La9%p_jJTt:jNR#g?w00Oin6A m#2]ێmv m3\+imiz6 Vlpt3iDp!KL&WA(1fUEoGM`rEue >jeM! e5"F5&z7o1vfpL7W6xlpC#0wil uk-lHJ2"S΍ߥ)5c@aTi~o;C.躁&c8X4ɖי&p:LH%ʑjMJ8a0FgeYL4ޞ:=3F 'D'Cҳ6;D@՗ð2u"3 M*jrI+/ei{mKi&O!xrpoW[{4vn݇iiهiWpz*> TUf u>+^H?8;wK/G-י&X4u?A9Jmq7hӏ˲K|2_φY$GOx\W"ځ5}ef{ŸҷٕUQzըA&0yA}-BN"\EPaaly:YnvJ(ImX vEg^~ XҀcIk5RÜi=9~׋<f܇)zL/J6NΏ_3#/*Qs\N9 ºS#?:¿pF0 ug'qg\IMmVJcż'\Q+ y? 5A\'ug*HwӾ$rTqYL1-M 730K޲lWvMf\&4#:LW1 !-+6XߜQcg7x)z;4 4@[ ZVIPV\Px&Ɏ7aLJlT U 0ӆլ>B]K@iޣ`cDv \ ĴkD.nsŠhݕx$O47B}plx679AsS`b2 ʢʃpW Ey)3sʑUXwxK.9c:Xe%)Ԏ`wԜs_hX^mjTɅv%m' b=6 \h}_iY L"`tTNFi\OGizlP4QM67@qIl1c2_B&Ah J&j>MSllzyVE~]1ٜa{$éX2)nn]Op|ϡ''c-cj8GBYȱC%>hC)*pЀAW'Jӫ~QMOom>)6-6Z-o{bُ;' !_J24_Y_l6 Cz]~,>O_eu(.H^0k^1ah:ƀ0O  eVɢ ?I/Ez?ye8ϡ GŒZ^]ބs.n?cl]Ci."vԹoc7)TqŻf?÷f>ջn /*C D:Є܅~t۳r.p~Hcwj[ !Ȧɗ o5ެY.]b-[zY"uVw7 |Nܽ6= UqR7/ cU?Q-Rb=|F2B*r/[9G`ZIJ>OABMkdZ˴[stF\;h#ZS%K.&o>.L9s0},pw~cLaX6ny9+dbd$0 d q o.?~(^ Cڷ_1໱q6˭t{%*Ċ$ F'2`ۅJ88G%&gęDl0IB$q4u)Mw}%MAVRY/TϘLgġB5`Ra;RT{b-GЍh`H.VyoG '[%.0pFe26KaֈDʍhQxSENww{t6R 22Af#d,AOJxō'$3$CuХmؙ3J%6T KRe.AOyJ6YXKrJENKPnC)ThfJ$]"mL1 SDja$æB!Ty %NR J0L,5\I)D,2Xa ҈0IΎ?֦aVڦ@SY*A!YL kuǂ[HѬh-ʚRbd) -UF,I&E!H$jJ KbR `0&t0W  p8lo@;]LY2͉SVC2q;O5i ͘r&=ZDI""8d@FфPz2@?A j=@)E+Y[By7( s.|^8,*VJX/'R% fcI;i=fؕ:tgi&0 ըH Pf)4ͫA*U9uYAދ*ʺ,WJtP_5פU C:[ihLb1^yV NiW># ìo.y\ӔI(Kd0umMW-d0L&Sas`*m˿kFŨպ[C^kpy(Y5Xp4vf c\u; j? l4 |殭@aRB^o*(9Q͡xy%\CD[9DKhLwPzTaH>Jk$=|P`Հz[ja^úec4B^"PD W}\A' W^!0Ftل(V2 -TM%h RŢ4tT#>j &_AU,:!aQ0'#48pci 372;kЪV gF0(I<(2SLGjJҸh ҵ=k$Gv \-vMg TfQW9F D@(ťpHJfH z΀zQreFpJ$nÈj|djO^S:((q˶ %.I%n% D…`~?XfИ 9$bUr "&b٥P#HlR>ECd؟ ESl$Ti ui*? tjwW$"Neg,2B- T GVͼ..b cYKN v;3a?k觟vu[̡3ڧNv+٨zoqy3_S"Sp)fmG-7gmvѭMկG1OZ=׺]'?,+\ju|}|̅=b_{vj g8[+) g_۫_ٙgrQG:dk`x:g8̨NxN:ߢQG7(uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:/֨혌:v<ϟkpQhqEurDFoȨc0:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCFkiPc28T01r9J:m)UQ[44%uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Q:d!uȨCF2Qu|Gs?O^4o;A[ewf7&s*%/l: Swo˪ߧzW&Wkxta25 pwV=oxL`Hr#* XnՌ6 fn,a XHz&~ޞ1 ض둀m;j XvHbV2>9Hr97SH:hz.V|5J!`-W|~IT w&wvKMasI s#䉵i>;l\j~=%GIR6+7RNo:rڳBI(eY 2_0#kQ-]v*1E=Ja;Mh;$1yÿȦcIٵ5RO47QI(jiqޟ-A cO?v SvXX%N_ct{zoft큈+njcr͓?sOv:X/ > s; =t9 noCoQ/t=XG8ێ8i7C^=W9D1pqD4o7 sRLwA|j՞-y$W <+0v=S-?"ŏm$?W o%H}o^­Ql;ol)M|k{{i3tnr6&^Oy5~yce_ǽkgР7Yfzإ/ݦF.O?çi!?Z\kT˞5~G 𤁇P5Fվ>wK}mKeOS$B|ڝ|J=ɗF׆0K?0o~PK˝KlRJf=QCSDuhq-N-Nƹ2L16Z)k8ES8u3vN`e9E/,(f{z*m_6|j麒_e*{+h{\h0p7@gw*E*&<] !Z2&BJdVM+$X{\"F k,U ;] ͹Tmu%/ R~Vn.Doa x{ĵ{05ݗWuGz`0MY (+Iꔃ8fJϋ>\0kP2;7Zzשּׁ08GWzFt;C}9~]zmmwLL sݨ|2qk?٥@~kˆkݻ$vf7ʼȔ=Vn%"iĈiA5 c9hF+x^|t+M 3S%.J:ÙW:t:9@hS!+ֆhJrVTyn+* }V m BePekUrC= ʳ{3/4xt@v \?/7_tz_RM%kin#1|?<{=w"2-JIJEtQ6ɞ(@De!O_!KɮO*b.ӉG89Ynu40KOOoPw` x oSv woWvlZP _>o2n$PʼnMD5׿0Eۂ SM`9)͚oc@6EJ!*ĚX)SQ! $bVBYm [o>m\RYxs\'mAKdTS9|5uͤOVsH *CQ_^(a4⡓s!ԚYbhU2X@KKAB̊x QɞAgzđOէ. RR2&1 A+R.d@A=ȧԳ ]!+ se.p'pFr"т-,qo 9a{-ɞnqukh="kOO`FCPA7:z_cp"dU?GOU9|o/ fCcE<<Mz?ŖЃ,$ !%,yr1䓕<a; c"DPT1HEZsR/H<@d.q%yAȧ~z|J_W| 3 ҎϦrM),m: 廕Q5Wy`\${!_y0чhMY_f 2>8GGOU9)8)<S}:/KR$ AKòqP9GO8*n6H ^ks@F'T䟷ɘxsd[5ɏqu=,n0o8WP!kM+6'sŖB[Q_*n,l,YO7Uz e5\=lx*cڐiC@ >$n`}t O]CDE|QAi!L`^KdNC:'h!=\?DOիgAvF$kQӆJ(9xCE֫qn(^:熖l>_p̗g%L{ Ǵ1¿apUpdseF$̓ZuVJ{ VK J|{5?īOUsEXyK\}\╍[ huk)-h)E+|\E.r,sPa`VA\]Ӽ^ yVo\قI,vsW_n~$u ŽhmD3(l.aԣȧ9_I h/CS2g 3-uŘyX.dis= ϙS]:>-`cI cJQՖVvr z(7{-ڶbm{1:\2Vl~|[HZjLv~OKZ|ƎbE"N~(eUlXjDyh|\.^SW QK ,(ڮ8Juu-68kQnl/vk+Й[*%9JUζ7%nAor 93mhv(~}H,|52~oy%`jSPEnٮbCo9PCq+gI bY)~V E>UV57 S1`uF3ʚȃ/ i/εVA 1 ֬"8;" PuDlE58?zra'jq#:U1՘ c48hBeAe$j>`٧GCk۞v..2<PSU>ǻP( c絨اzuXsdJ?|[Z\r~;ʵ俠?\_i}n֒wV$xvބq%'cxT#r^w/~aI~w`YǠ0FdHHDV u~LahxDE(9L2p@F9AiM3JFi8mD&Iޒ;Z51DWd?/#o*gշ˘vÀZ)Y818'\QwH#4 \"ڊдeLtX5. )tkX38KzygD~-]JhHLpKtDД\Ep`ՉX>TVCw˘\T6u`LnYTxpяE]}Ëⴙy1GCUEl:'J~dz?(*GlD&P-LEM:hg,Fl㣌" ̥Ktd66>6>TvVq-^V\ypC3ڻUWb2lwG8`8g8^a;2 hBPSUnf\c#c:d4Zm`0>.aq$~|z].6#ƒu- /*Har7I h9|<F,QAu h }\3bfS)|g C?ĕM{A' h|*_?>ޏcd4e|ΚFJ#&)h0q&gŋ2< C>U{Zhl1{1t6bhŌG5u_Ckz K9 /6tǫ"fgZe{T *XpeDkg4Y.ɡD"#0ڙ (_ek\||>Zqr+’DzrD.H ~e^wg>&b}M@v"P2NJɀIQ 9-d\Ř.DԼs OD0&D 9uRK>@ APnVeZQp=φcFV֌`V^jlk'vNҰf,jmp -(tUY,f$HFeHGRqxBGOT-|{AMmx|Q͗vØ :mJD& 0B%RR8~@~@OUyNFỆFRW_3 ߎZ3~ʿıfz7=~HtVzZL3uNǖAēATK41dohC B\b7n7{ΙsmxѶ-ܞ_r-WT;c<[01:|xs h&gn4pB禡EXX|l=|Xd:s_ߍi @c|Է"PM24ua(|Q瀌њO\j7>o@b@9Mȟ>{>"I^Ee+ &hT>[2 㘫FQUfMdc d4gh=.@W ] @mZ hpׄ妝LQi$N'JP. }e6#XD Շy1\K+*n^* ^^`M hx|rXuKh#~KQ+fg-8a~ zKpך 'L:,FOԶxa= #vmdJ,Bz+|A6bz r8\ ,#(!Pbe*Dem*)ɀcAOU-Qvckj}.*jc)eV G5@VŰGrY=/<i*;<~č`ENu-  |y#eK|55-ZTVlow~D`Q!9c]#R@OUyM<TGbC[6Ct8vFhX3 ܌48 u5Ou ѲA1v>x8by{@Fýܶ& zKaAI&y~Zc jԥ[RP {) 7E7UAD*-F?HyAQ2ZVTǧQsIJg K]⾛D/f#[ux"_kE[b[Lg;ThDskp w-M0dR8td ݳ7Edd*:(v{ |!:eR:g>%a: +1#:*G ?-:=kfʼZf|-V3h.i&uԍWe>Pu;,+_0^"mtATv6W~xq%#n䣩e:*FrvlI_ uvqD"£Ǹ$ژ5ddߠ:{`jcoe82@{pBAG# v{|5Y*"gsx7[/z iY,by?pc7w/l'. ?9)K⼵8zP$l)[ƻQ6hJ"Q:@}|bRՔ dĥ7kw?8JN%/H<_,gSʶ<<[ٸfeKx;,Tox4pf,^ji!t̥7#.6>}g|:9(Y5b1d'8Zg$M3+|kdx^Ŕo-u]v.IZ7._7A+tevXK o!C%4҅f+.|WR"Y( \JCzxt[YȇT㇥eF7߇ҥi9l+? Յj+DfA_G}MNVd1ԊNT?mȥ2vx}r\*Y~!m&/ʇuZ:qͱݛOWLǣRd3MZH̟%3 ҽ}\ ^ޗrmUsлn{gr}VU>u}.|WR:^qi`DwRR}\}٬nu4oUxSoɯxOYߧ|q6m~7h2f:KmJ$o>L4sAlg~I?OsfNJ; BkK3<[ :dqЋo7X= >`8~W3,Nм'zA-{A} J2ﮫAM^ezg@1ZX,(+b=R2Ѻgyǒ,SY9=ZN1Z0u( '7T[ Qڻoaj$fB82{y@|a@!C#ҳzPAVۣ"EggHD*RҚoRFI;dHGx(hI1^4 j+7b `O, :flǴ)ҴGɯ>j9Yoc*uB4mB_) p*BGݼJ>Jw1FGX}0l!! 5܇ Ĵ׎fƑSg+kb%2d~g ;?iC_NdBe^5/ xw!@ Sl'iYWtsa{_B!\x"IJdk~\b,I2dFiuMz(kEb҈1Ty̱B3VbO+]= 1mhr"j:VI;nCkYr$ \#MS N9Ĝ eCkKh]1ͫs ږbW{ Xg=DYiM0{ꈶE -I(tIe+b*P~mB 14@ͫ3hjAER /z2$H1=ڌK E':@7I!v X~25Eb r)0 XrϟB{ykբK7&oBGPm͵8VZUM5mC3MZB մChZ#-b 7@ )<o&򿖅e> B<E ї =Џ?Kf֋yi?)HA7g/zuvJ~wZlZH!oS {.R춬qll>0v`-P8g[ĩCb2#5%U!+ޝ2z{c {R)fdY"S)Wŀi'`E1O6b@h!Po-Pרu1DPe,jH ^hǮF-N+)F8ޜ|hCcD-E@ ygo9 N m~@:Oxru7yda(A"6{ZJ|!a'bE:(^‡6Ի_oDZ Z0BA\y MғB$|G/DA!"7 [1y˃$:DE$t%a!($WQSq,Иz2U=ò,]]qbՆ.U*\P%^  yi%hZi|.'4h/ibXM}\ 7Lj*+x%>czH@GQ5Gx~W.= @!|A0BA;O^O sC>ٳ[AH'7J~-$P[b jl3%C/;=ӑݔEr$(}dF(I*"m$Q$ZoJK d/x}dj;aKm)<;a!ov\2DE F |Hg {cw!"6WlQ#j+l0i ε dW;Y:nF B<'mf>ӳF[-(3ųT풹$t/Rʽ,i;lwuP~m/nQ) =G8\>kxu9JV2:#@l(45JBZri20_mHGD;k<%!@"וϠE<"45׵B3[IaB}h ^y8-rPPgcZKPck9hmM74+61ZwGkjLb:Cϻmp6[x$ztUXWK[{ i_˗H2|hQ2"y.I*`Ķ&u 4)t]LV՗֓v97ƅxևΦIy]vd5#A2Z1R2}-?5Ի_CF!8bmhi 5n1ATQO1h 2b[}u/8X% ]!ѐXdcS ZJ- aLkwP"~mB c NI7__nџ1u1H8D zMx]OCCQǘ2%"|q [W!Y=zwc!18v&iwnlQCG{@7tC{[l.gJM,?YzDz9?E)bT ^aO)OEU0@,o)G [z? 5fc+Iw=xoDokp6sJq Ӷit"~s޳0S@N@>|), iE43eo#V @iB/iNLC*jA=&A9yFqg-b}Aph60-)Q#E;rՖbm{٤0hMqjÜ?tCh?UoV ~Uv,Xu: ux&2Wp%IJ<6i9/BᵇD]9 +rڐhĕJ!>y uDjy妎vG{A{ޑ w-r:&71~AXs-S7 ٻ$B,IӴϯnwi6!1|֠|{(>+0 5z[A3$*+1cEݪ1.tކ 5/)/ZE0 kkv1FAE$ebIu*u6Xn ՄRoc]d6!7x(ȻD]"(7ʤx{?dgq^ݔ{6pzHe{0%%+h==$" IuAΨ/erd΄X*s?b΂$ cz_X L~iXLf1Le,lŲH=,Y%YޞyC;5nVh1`'f!l RwG06Z0; MtW(jW-%ɶuq$RIe(預cuH`` N4~AZ`Q?A[Uuq08ZXJmb,B 1A) ×OLi~'4ƺTa`i,"|ZrXWRzD4~j<%<4Q'"<h^T {J)q>R4%M)%B0sm]Ta " *HbCJEF0aT{`Rz|Y׆|giO,N!x3eL#_P A 2)辫|gךAf8l1ȝ)Y0$pNwf9!RſĎ/Y;?yXr+|{pFҶja0Y6PwqqD06Nd1ݵ7siTُ<#zpvW m.ۘ]t@۪xk巪ي xJu A ԩ8~`Ms.$ /[;oa_Oͷpzo1N[tof'PoEE1֛œ53m/z l_hyw_/`,wb58lEQs׍"򧫫Q.Gd< ,f;Ň#xKތSh׫KALɊo') :XBR`)S{-V5uبRY?Ivj=f ]eDϋKW˓d }/㳳@v;:zUs5 ʲ啑jU9(5$HܠTڥ=0 >|B&vw;t@? 4Kc!Ԑ!I}|H_ѶuDdEiEGb-$H]A ߠ91T]E#}D , U1r㩝;ֺFbg=~U_ ºjXXBB8njk$%yG;D ]>õ]N}yį##B>v޽gz-Z&ۈ)SEvLa^j`^=KϟӋp6ù +J0gunharvG4+^f0W '7,(w51U@xvZ%ߟmg08S[@{FN*2*tYōxޖ1qrf(/%,N}veqoVy̗ez:[,UU]}ei0QD7nMXN|+ĉῴAgy&.ms^Ͳi66^nBu9LU%缊.Iq?}mz__.Y#Ux%HicU:"Cho$E/A ƈA c/Hߎ샩;ooo FR q*~3lvtɴ z, I#4A)l)"s]BLUy: ֐\p-@a%V4vW%WJh+U+Sz)^U$Ջ@)ZS#a#]nզxam⥍OqJ4O)@4#_#}ŋR XQ%*~ ,tv)δ66 16 8Yl,΁7V  qP@RƈX~ʚkD1@k6Ttы"4UnH>;E3T;<=l"߰q"b:T&b1<^7(^ rq8S0A78cw!SEPjoaw(57"7-A nS[  9%] 7B U]РX%&p iz[O SvChQ\RUSH[6v[9ZFz`R{E}}w-^e@ P] zh@WVZJl19Won]<0p_XJgeIx\~u jezp\z@oTMieJwvFJ/wM(RL`8; àQ)kLcۮ3nwC8uЕ[5O9)W~>mP HM0!+b|r'k+8Loѩz{-׊`"tt162gJ yϣAE84-\J475W`aw.d`3.V[?R 46TJ1;ibTNgU߽#J;0oUՄ_mk,8QEOp-]˟w zTjNMoA-fS?[&廫'1"ytf^G?5TQ?Y](+Z;j1c|h4eQW:X#~jʇj:#؍PG,#`m,oW{K4}g r^w XHn,~BY1+B*krX֘K/&=;^?% ެ_}BcCErn%kZ{` C,F< G,Z8HSs\,AL9)A4DnAV:fwwiyK/MLZ[q3ݹiѣi밥!_mw-] a@jd5xO,P,rsZ ]>%u kpԱ̖yF`ЪX5fD~і%|+Bq[ 5ɼ4JI'!X67HfgO`|{O{Ϣݠ/NwHbq/FD:HO|ujE#=)YfktPBbx>  וtle|_p5fayn(pB/,#  [QW#T5l7r#+(KzGdP73LhAT./}TV doRoϤ]qFfHs1so S vuőAZUx*853z: AIqU %U-c]Цp({ge=fvͲ~lI:B#3s*~X]j?P=)ygsxr{;_U0(D X}LQt3+9Μkܡ̇ⅇk.a!pxW]W423G^Μ2mЬ"|OU$3N/W_t{Ff XMx M+/`c('D  `&J+g&jvniӪZQ @n⪪FA6@k#TOoJwa<<34"񙀂ȅu-/gFHg gf1!!Q !,IJYZ)M—͖%KsS CAQ"JI(CCׇƊL+$ܙ\ 1LOq*`3=]Gt \ #0(PR%kcUUj6}27hR#423G!t]Hjx-bMs nq**$T_~ue[i0*X9^W K:ӓvQoNi=6AoWqB * !DsQ@bDN&MtLP;de5% "9DuwwjJKqk)xH$y3E%k!q#@N!AZ'0#~QϷ3ieL]#CE6a9+y>Y<3@@`׸Qxz3c>Ff8іXtv]ΜU:sΏ2ӹrvv>>eίyWٙCn ps3RR6Uѣ$/RbރѴ#{1xuZ},{qfN(2\0ه"Bv9Lbͻp$1 ygUcb^\Yz4 C])zށ9n ?v(Mi15Jy++>m Jghdfb=IVbO{H`CJ?2geg^#g fnwPk)vs "yR*4ƕrZJtFfӔqv̡8)A.cJ\ͭU"]|Kz֝f>yri,>٥G\6!B $/QDR4}Iwff(&;[dTfViUmͿy,Aű$2ݳ?y6O}ߺGhdfNwv~UWpHfaRRrz^/y>UbK !tgJw: (c]t!$E;B#7sLON]S^z̀İO&| w;ݿJ~AIGhdfNCt> *0 qAZܶʁrŠF^?s:o%=GhdfXqo:֘(oYT=܀Dߠ!Z8LVSa43sF#\!CqlΰT$n $\;1˜)n!g`g'\wkR%iZ*cb%)45]2xq].Du6I?fpۣ_r~[nEH>ufni|XL4X =D^եVa2`ZqHYٴbiF/ͤ G]<jXD~_-\-O-g8\VbeJ=5= w-mHp5Y"$usl%qŖ\\&ŋ6Dz!)Ŏ$%ٔDʳwÖ@7@FP.g لʮX%Dpc`M:ߖW-Qp|[96a[ҵ:A`{tʀ2[ұ]YvĉHFq}0Gvi+) 00aNbp,T`/aN†mJXCMrWزPжGʯX(%D 'C ND!Ts$e+ ^dEt{X<^4OFe(XX@b-Ud^⫕ )JI'/ܖYq9ȗqpM!a%Hj[ڳLlI]W!ˉb XF XpsLe!Pq3} Au@\ *! wL I]\,P4*N 1Bk% !4* BNT:^i[6nW,ޅb#7#JȢBy!sZקKkl {A0ɦ~!0][NZVVBzKd(ֶF±i)YX$}ï0N*HDŢA}"Y@b_"[S`4*C^,L .-?Et k -%TĹ-Qp Sӆ5Fmx'GPQ4@kGL9 xQUsu1i[ַ pXn4@UY3LUJ݈@eGCi6_ m 3>  ~ ݩ G@, 3g?z՜DkteXs'HU[þޅ ?}5j!2YZܿ`:' \_sKTӽ ;g$]'ۦ r [+KY WJ2RPO]f_M>-SBΈiyfg~r Ȋy&Œ .M /PI(@YRb)yĮtO6'0Ta4N?HVZP8IխA,+ kt/$i$gjPO(&^w4o,|OA&? ΘvgL5]=s[Ε':{yٕ'ʔjR9!`Fٰ]`*$ ,iKd 3֜t];Q!qL=P:u3zj6K|m0VWI_ K42fեR阸ժm-fpw~0[{Bv`2ήxJ鳫` OG㽧8$j5^$ōzFoQV޷at?u&, c!B t-%@OӬq0Giq`p {Yewj}23LVJ ?K@۫WxdU @xluAmPIIJ@U% #_40L:1u-&*Ta )*l1y63SEx(oYl+R >TuMh 3Axj̭Xˆ>,+^T&aqC_@˗Lfϲ#lO0wj?^6?6oVГ_vӸ'PE0èu c@_A3}f9UEMYc@71[雳x촚IoZGŹ~^`PeG\{f'*;z࣊p{+xĶZ4]ࢣ23+M36WWk'܅1QCax&X M?u17k2L|c\ư c_On.R~A%2ۇ(4^]o,6^ַ2/x=_5RWVX*u2h8b4ۼA(&a8Q3)?`NƧL*pt_<3:o-I4~l8_NWe{uUzU?޽tJ\~"@J)̿1Vx<*U h V+h`jt1/g5~v+VP/P(}ud&a8_mqV`7/eYąVv5täK&S9Ko|euF ϤKF(}(J} r5 * 8;V%#1F]l$.yN?B߻_OSNSokQdLwӸ=Hy_VnRZsFx4o۴ȹizqS!.YnP O@{OVCYzWb1r׆)~ffsM3Ic-6$s_͕p%vu$aK{3G6%?u@>w؆Iș/Pc#A 5KhuE>i }.ky`H1=϶E|Ws}I-Ob)<*}ۮTq/.嫾7`棸gf㨿 j=Erx 5U8|.ald0ӭD2L {3ih)0pT+mjҁ*-ly(1a;N>8nvB ` 42ua&,8=i6;%& |a@HclN`v(Q0Pu%O=jYvd}u;>UR:lyz@Ns$^*Imm;:?(IV_[[~ː"}IdaF @Et 4*=wNsuC$z}v}϶,npk#(̅GU#b0 OF6N`YˁC]&؄r&V.TH!OPpq86\Ɗ Vy><+Eӹ\IUڠ\ m0 H13]0j TI)s,^:o=Z頦Nγm=ɓ;wzmv>ks%̫.0kVPA go4ONz4\ڛrŒB9au5[i 裄ä"z6>]ֱX5n]F9&nBDzfđ#lc02k ̓{{;i5ː\q fg;l$n9x[.6," oـf] w,,I$L 31-eyHpw e苵#jbM{A=4&x ejƑ_ai'{$HPU*xRqRdClmdQCq2k/$JD-T96E Fa( "ll` l o|(!WMN_o.f5m )=K=ÉK eL^ɯk\OWOuf2x)Uj^ZG Q&ovRVCC #`nB}O̶mȲXk3KmkUjWq,5 QUhJf*u!]/mfoG{,*) +\5ڜKo_+.@7p`⤩% oܾ}w(s4,Ce &e"wA2׆*E468BX FŁ(joUʴS]BT%1em:3Ҟ')m zi_?IJɐz:' i \+CPkJJˬc+)epvw-lj7)cM\q'`v;bWIɎ|1}\]ZWR\WRZj;\I)K ڸEp%M]Ii}p%,Sk:be OBؿ6_Ex/rzJt{&ZZn'wy&]pg.|R.Z}*Ȫ>]wBGVǷBG23{6g"AHC Qp( #A8 # vg<[Haf{}k}{]oZ>xk}\>Z}G_룯k}>Z}G_룯k}>Z}>]gl<}iLu6:뷦wkmMզeVt87x%qxoϥ?Ᏻo^<kSˣ;930y~V%N!c\Y dOa' he6)8Gջ("M2 Y9!҃݉ G&M<EV}p|DN?uY<?,LhUM}w4OU/qt=I%|fZGW@Y ӴljQMJN}2$"&al>5-6Jph{&дŦk2B T l(űpG ,߷À£>i@-;DZM.: ?_Vb2R^)@62&.s-^{ @1."l`[`i"?/,Cw W8{Ϫ^2r0M{]F$}1733=" |C@?zGi.y<$)tॱ T1 k33~aXd]kNF8>D*TT6TWףм:8e0-LQXEh2m5L3y8LS3ϺPAb! zS^t3b<ǝK)N*Ү_ٿ?}ϧӏIgP^: 2i.,NUiJ,'t VfNj-Z8\}p3_lȲ St r~g oA;*)yPg6z׿:Տa783R `ٻ~qv@t,-޽,,2"f8pp.wWgU5p+{S/].^ ZիOЖcS&2 8*rDqphq6^oca}e_qn[-:/Ƽ5G娿Љ Rr$h|ݙ=ٽ;wf?dJ 2&3w,S~ RХr?h\RB/B/H po҇ QH)'F2w˲}j",.ˀ)܊E;4y__ON^l6SbĝY=;?~}z?*G<{?ȟI~lHޜ>S$M(hQI-_x9`bU5l| z_(jqiV]3y1e^zW΃i27mx2/NIdr)ly%u?eJ+@. E=^t7(a^?C&6'd ꉍp<¬x6SLKeXlV@x!Ҿh_'ʍ;5驔ע߆NV`n8u'c~y݌/]$0&公+ThBgdAUbMK/O;]Ƥuv!Wt͗9[K[.ƂzM%+/E)p֥i{?!]U}Z:@PDq0e 4DCM %=ۍp]cKܙ5a(5a/,*z4- SXFƥT3rC/\;|0GuM~D%8M nZpj28ki RYknٿՀJ#'TyX9s/4hf^'4MI8ɩG'Tgv5K&0= s̽<'䚚jVZ=L@md}U{tפr-15ٕkĔ C,3 X92=@7eb= <-huע"jXsۂU+W;Kd[M!ǢE&k[I ES=H)ЩN!Syn%sw]ӭ&m-#ٙkM Y+)l٥Pq[!L ~qTCΞm1=AgOлa:{Ξ' a[:{f:{ΞΞ' da=A#n :A&1-InĠ31ONGĠiĠ31L :Ġ31L wL }L :bԙt&Cb͚ ۆ>O/kJs䧗[F>cCtO<ڏ̲M[O 8'-@Zc', ycWL&Rĺo2X6rS`TV-q9?sP09C=_]K=[T}1_+{2v>`{nh2yCMAe [v|bOE돌fp gx fpOUd:B NjtHΎx6DzFfRqٝ Ugܸz.cbxS!tnY2Et3ΌV%=Ʀ [I3u#oV=lV0 /™-H``.@b[@O2BwUI:fhC6/_[q-(0àW%5 *}ѫR:^尨3v"-bR`bEKq-eGϢ\'ʢ >R%}˚{UQjZZy*Oc\=1_@rrm_3Em՝MQ43}@"Okl^l;7;LSe6>:lT^Z(!ș L>3WNMJUD4=矦;u%4pn ;ʴč%㑑~dj6KM|7>4eOO`a*mx5\ZN.HRVC'&8eL,lw@Vp~R{ [^YM+D*5x3` &g2a < ʋ5~{ 4^$[xs(\hfea*s͍RBnj:M78X\ Qs7]A|ljLli|h%P j+ LKEX{˅eIR$:)/pz#!Q(I @&\Uem*)7H<& v͐RՌ(b X.0K{r!(Lu%W8k&K0%5*GKC$P o$4EĂ(?@B c88J8/{(+ЊiiࢇH O J|" 1+NDROn?DB -$ ) `2%J;ɜVC$P^I^;kH̳knR(> r za \%`SSrO"Z(~=wX9nl5kc=dLmh(TJk3$q,X˴9dWgyB4Z.>(*ƪ(Ḱc&'b-q,0=DB mǺb8D2A2)&.K  5gB)/~k$n5|Ζ8G = 5<$z7ra#ƞ g\-h1)_g|\Eէ 5)_r\ϡ!yOB^x+H(PCy:8鉦PHzU2BNz 5ƨv`+\3qAnqnٍP~(TGCb2H00&JeWP񅳕q[ 8˳pLd=Qz;a+/(4cD0b:zק VOQ$>j "JD(yn]I! 5g"!@hC)bYR^TYa>DB S15esȉ$MWC$P^QS{] )^@-}Gσ!%hM]PCyͫwB{IV)!Bq [0DB 卫P6@ IB9_aT% mOj[)k[>R'(i##H %z!j(Q#Bj&$$]SdX}s~*+̳ɚ9=!XU+&MgPAyAe /tkH'"8i#WH]z ^kt9%-a9Xq<4Hs|I^;))85 kRqJނ!Ws)ket.7tRc!Jsz1v:p+je*#+D POtux8]Nq3ҁ~{W3^dzӹ7g#P'S ZFd8%)z- 7g/~]+8K`Og/b | K|YbEeolf!3?Vf ctqg3?i^L@ C6~ތF:' o&蹖I|Q c ŗ׫xp7~ZCva#ʠvA.rrrf|3R^ތr6|= =+˷yr1],'> p8Vo.%]dʬHe#.r xJ<#Bd0!JLN %"!$^ AMm]٠w^zӽ.I%3H:}3{wuqjH'.5fKyʝ<^}ͳuMu8)ʏDѭUYaNh>JޯWUy~%u類@k$T\S[qO ~B/}ǧKoug͵ɏ&WՇƝm14;޶cfW78:UmUkʞFM3{JMAQ qοwU7J/Ze,brbOxPHF"ԕNʝ3*YS݇'qh>U!7g|k>*iE`%Ո^Y2 mbIRY$U}Q=u6>e(~3Q cT񝾎,(wŏݱP7~=~QXkl`-yB%4DJǾ Qӊc\ d- O]!\)[+DkF_Qӊχ@sj]A;O*EWr&RnVzhX2m=/?=eժpᚯ&_ '׫rj2]lOtW\uN2ZzVޟi6FF8ȒFcznc{8wW4{vϾlghCn#{ٷޣѧ汦g4DW 7CWWV >zDWGHWZXCti']'] '( v5]!Z9z]!]IjBBW>Q*bJSJ]`&)utt(9':B2TpR ᚡ++m+th;]!Jsc+K5! S "gWTwOBW:F[*jՎpo-g]!J~ʮ>2Mo)g0]w,tZLjr+]=ʜѦ\oQg׹۩L璗A;݁m *.mB1J\ eJP}YPjؒR)!8^Am^X&9Uzo=:nՏ[ n^QonHн]oξGoLWIt=^-po7sx󵏿~W 7dVD8mTd}Ӎ_=|nìWf)M:BEol f7H{cZ`'^ͳm.xDKRS?YFv ?Z ]ZG Q]#] ,e BRBW֍>nlNtIJR.:ijթ/9%$_49%ND\9}D[o=E㉜[LCtm;tpkh蟔"J}J􏑮V?y"]!\LѷQR':B2BP]uhqdWVDWGHWV Cj,mh&BJ>1ҕNR#`]!\Lvh5}QmS҉> ]ij4RCWWf 9eWGAWNqF~UVmN[LV|5mzqѝzkoNޜ͛7k^v|En8гϊnϯ^M:]3 /.f/}ȷJw/0f)ğ[Pn/b9l.4r(/;6.TJtqp;[֮J|t]37"vMNa[OkdĿWTl|}M;T0%@$(?FG =p8/>blv3){yF8BL/Ami;y]8~I_5y/)[\eXmBpeKfڊK<0Ԃ̂,jULG慄?Yk+|?,א)?>B:V}|[D૟!j6'ӫ+ųHcE''-nihp^ ydq*YR*ͷl4` !Qrŕ!$Ebh>>/v7LqJ*Fw\Y)I Mx&厹d SF*I'uRt0d%pMNԵHMI1!KutL:RȮPB@&тw_KͲnG0L3RQ/3Hk\4`)Q[ݫ N6h N@k%2f]K68U& ' %jɭZ<ĪuUJ|YZ4/nލbF\*"uQ44u ULl^vH')"/U C9Zoow*ՂG{{]`!I| xḰK< *}r*&{H\!iZUFB1L: Ρ&q֣2E ) >@M&jZ5 >xmBV5 U8xѴ<@= KH`e YLZ[]hQx $nGep6*YȩՏDC^jۆnbЗb$S`U0~Hv?n3|.N&!z*`Yed-B&ȈPGA]Jr AJH 1 0 zv5t1M=`JI%KcEXA%DEWHPl5C펁5$BjFP,f,Z{6=^ Т td!.h]q6w$ fj2RA ٕ ?AjDLG{gUEתE¤2,,HcF8&gو.PBЌmۉ6BkM{JZ!td14M]JЙD[vTjj5]ZVE.> 2ԿIIX@it_tВt/UK mrytXʌؓUZADy آsE5G?Tڍ mMד&%\I"uA;(Xf4hTUڂDez [bcW,XB|⏽^PD4cҠOn F Eɟt7(V1^8R XQԌbwH ;^AN l?dN +ZOQ!,iJ C\1Fnk.NCqކ`ݤ1jEr7 " J.YxЬ*7#Ʃ -(`jI@Bv6B!hL ΃rt I0H5ozF}~/n^6[{,|((=ZfDcg4_ Yl,Ad:_~9 p7{ڛX|R.ANKgg߮V?,/f@? -7B~7}g,}n6W$=Ih~E/puZ5}R=MX'p/uO0=ԜluA6Q>u¯CcuبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6QЌ:QI1''xdF5q6F :]°Q1=:ᘔcuبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6QgZN]+ݨ(ly\o?QQhuR2Guu>ڨ VبFOf i*/Ȩc4OΧ uahبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauϨ<✌:$|:+l:@u6DlauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q:lauبF6Q;ߝ|Gq}XwW<=+_팉QZ6k.a_+{a]mNh޵Nb5mcbEh@Wz.([?-[>z/n  Ǜ%iY}mmw £ʾw;nbYj!Q럗R Ӌ tS"nt_h `Rtd}hwsWaX6XKͰ8OXGFشoʒUS[3H~GK̆sGPn{N}<NZt(RLWk祫ǡPm=tS/]Ŧ+΅I}t~trJc]уh ]Ͻv8kq(mЕN{7#`#l*?" J^ ]e]Ri }O+Bx%ҕ[?# 8s+Bkdu"E]z6tEp ]ZvLW/|tV6Yjs+Bġe`zt=n/P/r{Zڬ._/Z][0p.~,ADSt=5|IՒD "?ZW/>m}O/D㷫"m>>Ge~s!Re!(t#[hd8zt3?8JNǟOPqvvc|֫}Fv}h;ӞYigi.7M׺Ϋ}LL>ħE 9hynOg> #9xgΏ=~E%+g(j[Ş_p}Wrl2"7,}N@}Xeʠ%AHkNuˆϫ@DB. Z[_o_]G6QQllcwrt\_VWp~T"}xoFE8t\Q}81(̈Jh}NXuZQJ -şV?-oU{sЛKSf~ojqjw#:Q}OΨ=Yb#]6W'S/ygHŪ$ï?tdr'NLKwM8thG@?{W۶m Q<`wBy? 4} ТEb6F0$YR%Il-ZXlQᙙ;gνs9=c7_fotRup%NcNjw&@j9u-({Jڢ&M*SS.E44J&(4R&y)S[t~|8<ϚJzcâc- MCrj*l7Jk6nXclH!Z{M%=X"Y(IOֶ> PRh>'NvzNz"z͚Ց$ WwhqvoҮ8ٝ{;zTkl0'׬1ؾ)IzѝlzV@^ޣe'ﴦEʴ{,'oR":}b].tHV9sokvu=o\.)?REn^.}߽{np ?/>g'hCWkcᙏkktq_/gY򪤥-?%OQQKju f%~.erPL/4 eMbTL:ߛ!?} |.o^4ԟ]FÂr~ۺxΦ˂KXjklk¹*uqL!̍ZK- U*X^{ͪ'xt~vFhG'xȘKrM'Gk%Shn3㌷J/Aǵ +>(7Gb _KǤ:'*I4 _sSHƅ<16,bgΥWgjTXiYu0K>-Yj֥ MKNQ\! zWCP0_z9{_*1 LTrYmHi S+|)gZ2Iezl)/=\􍛜 FζtZ&eBsY $d2{L$Z )7"w*w9$q:[j&.bGۜ+aZyRG "չ6T8+djTÈW\BG`+A~25߳h0`5K)3I]¤,>aDω |4J? Q=/%AVZf͹JX7T|'${ %.7`]q1i竻!ŕKw7wI'"LFVDp$Irzc zn:WKo:7 )6.%'0YHbHg3\9?J} hE~XfRulG.<rƮ y+A~k^^I{~6U?9 l6\jV ]./U?a)xE}! "Z=q\^zv9UUmeb}Aeٶܮڶ(SpD> ;ï!۞d' IWö>F1Ucy 1QQ2gpe3W=Sd{32rSmu5S>K0oG"p=l6>=a-7cZnP6kYXX^o|W|}?o~o߃}߯޿%:/"l^8G߫j<7hx܈]wx`MVys9s .bq~p0?MGWʍ.4i;Z͚0W6 r?yooB_zUc@CHjbT,mvGiu-O<ۚ T>FEz%L/ab٫,eQy$ ՂA kJXP~4KE4f{CY.*5U7kܨvBݚ?@j I'gjg/Nton͎&;f~l- )TdJ}6[Iy01ԂzpPJ9n )If,PGsx-1RvP :%&٪.ID2vr+y)cQD F)>(t-pQ }Ћo## woۺ_O-o Qov Ql+Q ӳۯ4]`OwU~,kŴ^ܕkZWEKElh t (BkmvfIyh56DXx3+J߼YBʝ `<\DMB4T4˄9h^0Ri͙Brc.'d*Rr'KY'i4 '^T+se1.MMJzf, S߆{}M,.!J"s-׭ꯋ o6cCN`> Jkt3,$rZ+_6\ݛv{Jszn>)s/:y.2Dҳy+gG_R;Nj|y\sC J?n"p%씄hlNIRn)R΍|p"4IncX?"3&8ʼn6kNY &ud_ۧ&^0zN?A{eShɑZC6k:!Xt9Hco {V/05:Ry%Ƨ\mUτpzܞQdtS)ra U[k4ϳN6yAkz>!?r~R"F+TD8 rEB֘+#$W N UDn=]!ʶ%vtuRR"+lYhy QJ ҕ=@h)Bx!Bt(Uvqte`DW6EW ]!Z,1DeGW'HWVjmH@t "B۱YHCc;jreGz(#AWGw\?(2.H dyuvꆸ{>f=hrsGh4 {^ߩ7{|ܬ)'*XM Supϋh?_;MޟV ZQBI띎jՕmeP??g{U~<0="cn(>P\3$'pVeÿ.R&ޫg&]6ѳ9^j˻xb I `-h ƣ{4Σ9A \& ABW䶣+$ BCWW{5ZNW\YbMGW'HWBSix@t'+D+[OWR鎮NQ+Y$;CW ]!Z]Ji]64B6grNWʎN!D` `t(mxte`<$BLBWְLututeBDWXpV.wV]%?`{샓k5ȻC+tC=Jx ]=J kN+{Dzhk;]!J::Ab+|DZp)et(ۖAJ%!"BڧbjBS+I14 ` ` ;ҐNч}m_B2B>گc>U !]"]iJY ĬǹԂˏN= QnePIXH+lÉ]\Ji(thn;]!J;:AhR [ ]\ =+P|>t%vz#wU\IN=:,JݲWE ]=5+{Uڸwq#2n#7"eD*eմz[~@B5 i2% ȣ`<+T( 5 ThNѣa+, լt( thq..g}+DIGW'HWKu@t  d]"]In5t(ۖAJ ]`}=&+i(th۟<(UN "`N`rm} PJҩS+#j.(kj+˂qh=]!J;:A} kN r ]!ZzgQ.ЕܱeHcWkqm5G>۳ZqJٲؕAWv=Ui]gj۶_dv;Pg:fyɾ6^H6$ShSmjppqqp.v\&BW`NW復Pѕug+t>֖ۡԺ+*1b]+XЕxWʡ{OW#W.]`NHgQW-`СܷzWqW'_4z߿?-ssLpΆW/\,ԡ<^_O=pXk~`7VRLJߖNvϓp HPHGThÈ6R/aqq$,Q aM² lbp3-47O*hg>Ln֚e( ;闬@WNM/}C[ fGŕ(*p6H*K|)%%p|*e*?seʗ)XGqѵ74דxlyky\i]s쳺&&6-c-YYG*xD+̼ϣܛ'-|w$$NW ĸJKv3Jl{=$٦mxἧ,yR$dTťg#?=Q#i0#k fA"hdqXR)8[l uZѤZ`/itɬًaN 6@>ٹ!\!+¥|-&EI0! dPʭ,Zb"H8pR篒Su?uxG98RVr*Zן.,:Oй_zʯ/|b_6qޘ|Ua同BV|V*%h44,Iݺ= 3ÚjYK̡]%yH6g(y0p R"4*N `"6zV؈@@_ ~8g]\Q3+ytySD,!B8*@XVc -hVPZdܶp(p6e6gx7ֵ51Q Ml,pK0ZD@0"J c%DYnkjҐx031bn"E2#-<-&qߐmHV(1K:_F ἸQv~S3GHgcJO]qAb `>GP 9@ y}~.llg>6v L=C65}z`2}N1|Xl\{u 'QN+ignp ` tB"2)(160Yf' ˁ';J@0P؄::-.C7._[AF+5yM1B:&bh!Ï75,a"d:^)4 .1jA)riEOs+Ρ K)USoLqŒbō|q*6<;31_FI<}8/mKm+Ej#IVP]N&uِlTbd(dui2dP8H͹. Yk]Y2̆]1);U3;J QV.uITO+%m/M{?t'g_=;|mào;hײ-me -5ZY[Vk|X9T[pշ S$z1.=N'1띛̃nۻ5!*9:h |Rη2/\$Q**e 'QA! 4݄@ݸp}uÁ7HjbCTLin@+g3")?c!W;E S+1(nb2*hi!(&Okg&;ws?_:HF8beUbL7!60@.R&j$Vmݱ2U),%wvJ(ږ 4oPpC׮klԀOW Wb// ATb Wi\$V#PQuKraKeđxreHKFn<%A[8');I晍A.R݇Pcڨ&_կ c;upbW/5XNtښ Mؼ2|DPݰ:^ёVߺr[W8J{Pݖ1ݾf %eq` U  Ri1 d~^]GF! c};c*m'`$HM"`j=t>цz4tym>̋lmWẪ/]1I?'5a4Dq=E]َ; #|lUBsPZhM.KPT~׭k9`&AYd [x1tWVj_-U[f!D'l6xտ)hcxZgP?#"7yk]βܠywN%邤TQĤQ#0E4֑p8%bbJ&($6 fGItċhq~e,2\}t6)f={Ώr/MB̝qs& ]DR"ﭭ歶I.6;CΠn_ Vziխ/mNÓf2Oޖ{sȷ!h;F!y8\nn}GnH9q- Y 7(}9~O[+;7xͻ +&#.a5~wD6pn_pfew\&v_|x}nn,0k7~CkYy?vY]Uvx,^[;*9̱u|6a J@rK|FQkə*E@h 5̷ ?cw.bl^CFt٬x8Tph;_|3ng>'}ưP (E}fP& >[^tftWôQ.9&rRmlbiaH8Z)$$Gݮ"檣3uRn3vha,>౲?6ˇ^mttc*`K~xX2Gc7ط> h?&!#y{-󯴴S.Bc\+RFK!3[k,8Ɯ4RL# ѨJ8Ua;o\o`~OZH.7]@^_ρ[\UlՁ=|6+1)9;9g2DrΎCKPCƴ=}+rPpOWCWCt6UC3tΨ+}+gD"]i-ewywΈvp;؉Z:?#ۡ+ѰIL+|4.(LXbh7]*=CuibwopN0w m^ &%G3碑b0X##3ϼh:SKeUwFDېԻW/~h4?S-sr+t+e6Qs'bb=:ֽǂ=~=6y΋>zD`F#G4eq#ЊGhCY2؋Flш>kՃ1+wl!rhwr(%"N]9;}`'rhw T0gݡ+WtZ tut1%ۣ+X}X]9;]9tut%0׸KcWB+ANt%@OWeOWIW+d ԝ` Ztu8tъV J;CWnw28bՕC)qOWHW0ʻDW.&]+=]9j?HU6b+:*ԩwlsyt:*9w <8>yb6,xUya8vmG <9?ƥX\,!T9@DRFY,;n|%bۡՏĬ[ܳUngֻV=Q!LwZJ{:@"0;DW0G+WЕC+ӕCDOWHW2+ZiwAwZJHWJh K;CW.a]+}+8kXv\ڙ+=]9\tut%w*] ulx;dʡ+Ʉ]jwUw\uh%+VSOWCW)ź4 ;Jw~gCiOWHW TeE+ ,L'4)r NPb _~-z}GGOqL!\ċ\"͢4> !̷QhU(F}.}yݗ|٨0hRǧ&? [dyfOFZ`/0 i}45VR CګG=٥:7Wjcu2~gY 6MkX~ꊙdT=%^aDk>~MkfkJ5p;W [*1!&t*|FGT]V?jB/7*6Jp]VW[b38 rwU;QrTTʓj)ȣIw_鿟,vxJܛ@ ӛcR.ia1f?x{bʅ>Zh-Ȟ\c_nʮ}jE#h$ޑw Zvr87z֑/oϽ ]NKcA.2lT9yEAhq|)o6S$p䥁S^=<3ؓE|,.E*uEyR] [;Hb+WtnJ~SMWV!gi7=(oZwWAH.R{ AoP6n.y zKZ 7:ʑSeQգ>cʹ);V) a{'mlopd`y~=Y:tc[:6{zq2ZKx58/_;BpWWWVxuKb#~PZUwh=zo~bT+{|,DfW].=⍍B 9aZ.l__M\mj8b'+-4gKqbnAo\J3^J2)8?>>nL_nzw/'ta?|}Wo].}l-r uEw#ؔч׿_S1!ꈦYV 2pbd|)^~rx?D~ޘkuhͳ7sG1׿6^֟JWkeՀk@s2]}tT 8]~OW@AW|8)}o {g/8o 8`{W''I>gy]\ׯ__h9ioQ9ǩ5*kk{W2~Wut DoW|ryI>zvRoWy>Ysm~9!3M9[xgM+Gk1n/&%{O-}{ nweJY`{6GLm~w~I誻߸Q _Y>11Cy߽1,yqU_||>>;ya8~ɑ#~{}q-rY? XZwb{>ZN)ڗDުt^WﺏY%<cŷ:gӻ%$Ȣ>e}v݋oHؿ~ubzy8>\RIwωZ%7URvpS鬜%Q٥輹1]z%$Sr!UJB*^g\ȥlM85?85tZ}zW:P2Sh]1 B%̹I:q(@o]vԸ3a@$ZJQ+ω6[Cb:Ƙ%[b2&j48mg{IhSBJsq I5UK;WCwJ+$X2>^j&t7h!QsӰ=pCw;v4F~Y(MVN|"0`)dw6PQAm>+h-a}Ǐ:mZqDg()f-%JJcZuxC Ők'G(mlLBHY`(. -a54GV1tx.5R l[ W, ؆6 V4oztC^㰆*TT@5ݛsc((ʡ1fܳu ,bl85[eՠR5rsU #" +lBCI6-a\Ҝ`AV5hM鬻-P_Udː|=cf2 V޶PB]:=fBoTzFp2nPư)0tGqrn RJYdc̠JUrF`ۀ9Q;L`QO.F;/SzXX^7 LC6=&V&2x:p4@.A0W%LCe_̃Wwx~ 擛ѽ0Iv# C^,4zVP6 `]} roޮ7̰l+0'+MksJ5'7 #ud+1 b^8R.1UxJ#5Hԃp]/xp`\Sc,iLǤZ ڊ[M+qc Rӊ{Q63TUTk>@׶䠧G[>?5ع̸|Ǎ7΁ S묁X*C6(#< VAP4cVDa2Ypy/zWf3 M F;,`=Q8I= F2(8q˱ p3f @oޢp \8 l\WN˥1 X`XKEw,*f-WGR!6OS,)`f@BG`].?t=+{G}#W޲G]eVϧ}y,V0a Cq1ng/<a,o|sov*$w.}w!Gjl/r~=)| vZ/ΡAv|r2vtʛ)J͏oNw/vs O~{uFݟ+R*MPG[A`$%u`lu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPuLBݧ{?A vo:@ALQ:_bP$ HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑumPTL{:\R{ZG>^>Q @w-A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑγ |;.~:vs~o7_wsmWm~q~BF=-}qyfrEkƽ2!Wᎍ䊾\SOش7t5Rhyt5P-tҕ1Gt|RWImA?wʛK qO+Soȍiv<;_xtY(ϖ|8zux}5N'(WC٦"*Բ]kkej1f5t~ =W Z:'PRRka60T0)3*9 i`,/ϙ%97VP nTE֬јVǕiCYOei**扔g^) ,ǧk/Эk"D&7i}{H ^]ũIW <ݕnzp-_peM~hww/2GE4,صݽ=_j8}īj|2L/HrtҴp\T0j,E&3$Q2LFnjEJ۲l{lE@:tGy˦,{&r|Oo =8\_.vNgܲtV(x:K*n,GwLÏ׋LlCl>9Xls Hp'vu0&Ѵ٭NW=<_G-x d>]ǩwrދbw^[ g?ܿ#0ӝƑ_J\Qzvi>&Swz}eGphYKCt\*Ο̦:r?J鉑E.M޻*WNR30JQpܙLZyUxU='懶<Ʌ˥%<3!O6smQmuohm0?9%2Vy<ƳDkO2L5~̵Lmp4G^Oƙ_hWt1yA#by (jKvXuʥmZc j1%yT<q)HӔ ҬNc{K0Evs+)zF Vlz}xy8mw&mQ0O 5?_4:qG÷Fw4~Y6wuw7RLq3e:&<(t'0 2(ωL soDҶ`;;!x:t *q>ݟ*f)eKE gLzaFD(ad1"#.boOyz53d&]g .̼]}&K&<͔Z:i=&Ó,iW$mzeRnsnu{~6q2~Sq! Sj* Av3aRUN~6O#gmO/of^Ս-bTv%*<,n]mQ]fV_]rGNIX[$']z E;ΗQPX&/ˊ9)",䮧*ܮ:E_y:wǓo}3~tX3:}w87]ѽ ׃|x} ooysë~_?o.½xWo~z oA |U;%%vNpxkGc=|47bGczQ{|o5uOl8FȞn;ݫsSoI.E-h [G1݃Z\D*eqR aR[l5J€-Wfggg1bm pγDV%6uEˆ,ʄ1YovկH?bbƅoXn0sw4vCH&r?44^,yrJiT+M3OJ-cfV{y[lok#rל:9YJCSL(ʝQnuZs&)G *eT rMg`dܥN|1}_$X ve5OYy~_:qɔ}Sk^{ > fߢ4պn_䁰Se1ڰb7%[-:N9] r29]Ti)}md  PWV UF\W0NC:[ @8i+T+dq*WvϦg#㪙`y#vqL-ઙJa+W6MO-cB+,W,\ڶ&T*"z+f5Q4 \1\f:\z Uh]WZKBP j Ż+P !$R`p% ʄ+P˵:@"[Up`#B`prM0 U֠q\ib`64jG+T:2*MUqe(:$ Kbʥ6\Z W}ĕʐV-`pri0AT;2*U]=\ _ rFBm˾fr%i7ږDu%_qDu GCfLeϫyZbt3J™q'wqJ(g.zT|uPj>-oPLu .J9PlR8kL? _TV˾4{3xӒY̳Dp MrcX*e"^S mkTnd>RR =Ҭ,02HDNlB,˴72:ri3eI xQj9hP0hPICѠZ>AMg4{H+W(PpjYq*!843H P. ƺBLtWˈJ0ł 66\\kBݷP%W=ĕdj@|?"P. W U qC\)"@¦gBbۿR濌׃_?yt< |x=4$ $/F\v}WNTV%dIf,!EgN$gOg7ߧ1MI$I)UJ)&1HiɈ)ϩ)\MHN Q`r֒sT*J#{}ͥ5& \iNXBV+PKDBE\W[-CD<5 T+:*e\b#25 x8[rP ժ U%MI; n}\FZ`kvpL%VvAF\= i6ڛmLS&?W҈>IM:2H0:\\rfj]T"⪏h+W0yU ҖoXu\J޵LW'PBr@ `+THq* !2Z4!Pn4Tkmq*G#E\WJ3B<U ZyߕҜPqC\i wYx5[?гBAT)iUqe +c(]\dR W5ѐDQ rCm*z6b{6=ע;~3yU.^;h"={68#Bhھ`vT{3\5m6SٵǢDգS2Ps{4PKVvd^"]QkJ+-txt8؎^]1 $f3=A-.ЕVӕR]D"n&0gЕW"[JIKO."JHJMt$2hq/]Υrv%%n3JJMtFJ ji :]i)󮮇XMYQGv#4Wo4!5O糚 b u >B#dt/#4?,ލG%ۼi!a_*EZ  寷4XK5^8EzCf,YB9"bkJK{xͤd + (@@3BWZZNW +6ѕMĕBWZZBNWZJHW:xxWJ\vI JK\%rb])/L\m+-cWZʡrt ]q64An ]iq ki7JJ HWl+-DЕ^F"0P1GW/$͠{JkO]Kw.~=tkBo.!F#tij&Kb5](LJ1ȼ]qGWBv* J ̅5t# #Қ>hOWZJˡ+WK)-+Q<+-.ЕˡӕR:zt O?"c^IU=ĩ7'ޕFqdɂcbJm'8̗i!:])9V3_!X bv_t9-/z۲_U~a^߄iNJPn ?izo%4Ƞ^\ u6)J}Uzhc}3:J/m'R KHZ WGrrʼn;g4k5iuΝ4jy/ Zv>|O.EQ)~rUa1~o{AYcƹo^N%::mFÇgU)F~—6CE0#|=]L L>݉|3gY+)} }ӟa+pg|}k%1GBpKH:sU >Q?ΏZ9.A h#?j(9N'rLrFS..{%%NW5=t9S_>~Y&Uױ E蛟OyZi7嫻e{+?{"pO"#'`=Aoo :,$;-Btۗٯ*/%z0]tӦćmѬa| "Pr3/R]sxՃHWZB,+WaBB`xRy"MDG #Dwh~^7 #&{e@O")ʱUm*n:W_v`J.__rM 8{_'b~/UˆmR&o]Zx:_ yG=GK!dzGw [}jRMxYܱ6|&JأS1#Ͽ[{HWʀaţ2E{Ն: %l:t{iJsG''.#8Oaq "$I2K 1Gx>tJHP"ُ]\Qx˲|:\}?PU[&z~䩯;+jUc_uobѿ7_y 12wߏ+MMc$Tr؟n4D?vTO47U`  ^G|{> R'(I(f"]z Z& kTR&Ae+_jNAALXЄDG{$%< ?AF uAb?UGa0PC!SN:jxqXOd"6n ?5)xlךa BCT:p="}ֲ9p 7嚹/i2ܽtiAu c6cpC=|s<|3}q[l=oO6ZU0LNb &\F]raCN}ǝz8kd}3!_V6#a ڧ2pe}5ش2C;.p~!3x4wIS.aiZ' ύ`t^͇8y50#GFc`>YTvgێ=fmxd7DΞZǼx,R^5\r\;Ujm1Rs6ZC0|᜝GcˠϺf>ݙNK6e# g[k@twmD ,MǖBa ,rgQ1@$8@@`^e@XHnZYTM f (Y@yU{RGK[ih-%qU{^bB-C*?|]?,ލG#T+./h6{ NJ !U{k鞄G>2BP)wOgE?PLQ/`X'ڞBMV"Q`*/K]|ϣ"3Bt$)I@=7~H8V{wu9 ϣz"d,)J$'_ZMWܵgԫTVuKa0]> d.ցB@i<];ϫi ȩ{q/1t[69Ta+]&?zu V))CiR`PUR3f 5X 3㾸:uʭ{EWS!ѕNL= i =nPwU={y=(5CGqT/؅F#* >nb$r9yD;]Xz tuܯ: F8mVv'>*y/{qj& )7"]>XFoEbw}bӶpM3ûCX 2zqc}$T?1~%wdJeV`,/qNT.m7gO)^`÷(5뀃̤@i՛H$si<:G@#'9׹`9 Z Fp>϶]+1f^Vv.M]1\ +0@3o,|v9ϖXѕuƶ0}ʵ47v݇53xȻgOIؠ|F~kf4eTguI_my/@ΤcJ'>p&tiJ{XzHM{4!Wc*G3Tn]go@m} GK+v4gHUG)V4q|TiZpb BMP j [*2! +ݹ5B`QX-0 nKX%/[H)\X,21XZJO3o5lj8cފc֢,e V\n3UdT1qFxZkn}36hrlN [/G a|-m ^uPF ѓ!R]Ot/kc1 t)k[OtR?5پorUiJwfD!p<$$ x#15"+K{+r9{So6O>6R ~/OK~imC+Imϣ@xBRz(Dlr*|^>iLauwJ{>d5:u֋# Ϭ *ux.3}Y׻Z+ҮCqM.R{y\M8@ӏVaK( 4?w%eO_(uGmg_ͷBZlû~_Q1=0d3x S5`~}JLD"F(T$# aŜ7O9d@(A@c##x8D"pz4kף2tL1S@mGGSdRj-_zvO^9Jhi0㥊p״TL_|6h3CZe+XMq5Xx߾^kǏnz S Oa^:|q/Aae+Os $iw~$b>1WE (A Y$a0iPƾN}jb` q@ywk@nۺP)iQGmU]b*4yfLX#/m& A? M&`"%0@!U3 ƌETye(8cDŽb_HFju1ԈqP5+bHk2m;?> (G L5=Zʱ3z=GKY2z^NF#Kx(w1E&w{':w{'ު1 VHDR 1c $ )g2ad*˥O%2!j2m;?{vsrZ\JJӝSw%pPApģ98 ܑ;8=#w #+.,(K%{%P7М=vxk0dq7C@ZN"($%,!8_nVm?{[4ڳWH4Z t괈U{By'`DG$QNb2ugh|()br7-u$) ]q:tXv6zn4 ngQk?v4F5g@*hZX :-WE4b4MWȱD4Zi=)P""7є~IO7pU= NGpJ}KTJKT1)eKT|A5|*?%j4ZE}^Zv5}Nߢ&D1+X@1Pꪀ@@+UR⤮ޠQyW`eF f4~ L*PrTWHrDW:E]h9tuUvRWoP])Iv|6lj1K)M&r@Q'DH Ss(]-:v< Y%$k,f1Nƥ#0X@eCL3HDÜ TO#־w=zU7CCE*XbI-/.~I3ʖV(_aނN>p%@',`,0*^&\1~q}1F x}x$K}IGjY¬`U胬 ,IĬVeǕu6E N46ZA$"shJY-@Dܻj by,Bd/ڐO;xitQ*;3[S*T*FQH! s60i>,H?j_l6oe@|NUfTeNUfT: Z I0RWBRr6.3rOi>FӋhˆd;k  `!2%LCvH4oG_o>_t|¯~ m?:źroɢpZOb kxR/(>~X*K'fk"zO? pZ E5/HGD}tibd\LdT=鑬Y+x$ #Rpk6D&9h b.mdD1S Ť #!:s׻!_&@LHm_K&:^ Hfw㬒0qvݶ!Lx\0"_F1a"a/&k>mA9YrsWC 0[yrt2:$Oe.eBm g\r!|'XE%GL+]J[v#1GslٛS}_qڶ;0(s7royQ[c9gd\YӚkJ,^\hx(SOaq Pr2\cOA3$|XSdq*d,js}LSC Ҥ!$kKzb@κD/K=hM"X,KX|"D! K?{Rƨ/5K~r($!7X;DD=wo~y^YQ;ܷ|j@FYq|;)i닫syF=클b=_GxEhCUR5`)F B3_ ȐC<]F[Z9;rf/1lhK| ӓҧ('Ow ΀SF-$aS. ɴ{ѻЇL eHA kY溈U -%Ĥ_1hx1A: SzLMwF":ȖȎ1 ^|kENNx$߉ !8iWdN;Q"I.)@ϥQayc 97IܝVGP{(e*'m39%k% dTZpf(s+P>!eK]sx 7lha' fDiIIOHX$;B@#K˔D'nu#XCo(8oL7BP3 WJf^w*uUd<ۣ.nYi,Vzrt_ț}]x뤥Iȟ{/'}؁r_,{ZPq}޴n{7Y"0A72;F{f;5Ur!VXAFϗjaټH-ڔ#nKǃ}RnCej%ZvI"W|ySuOw' WelsH׮^ϫtb?΍dulѻ^Q﫫mIR[&SX9~mڜ?V[ޱZt~X=ɱN;(՘ީTeegg*` ';y^8p-,[F 9me<ܥCf*jC-Vvxx"܉juz>պkpTZa9鍺:)4uHDm#g#uh퐈7Nb!-|WM7FwfZG[tmXPߝ<ԇF }iR,1I1\376ȤliK2MWV%6|QGLM٨P3+v=ExQz.0LXv46 WOqʥuh@ց66:p&@*/OƸ;M7{k[6UiChy d /))r<~hgȨmP:1YY SQ^ohoq40CV˕BZB&%W(c'"ɹIzBVمtP{;F1/|bN$,pd|0.' ~#>$@8갏LK4'`,v7 nwmͩHtFha:w&b"fg_;̀J`LU"%a9} Rf}RT*wA{b1~ e|=4 esCV/U Bjɤ._.&T`Uӹ;HxkXTk@"2qVD?Yab_fQ.{|geúbiCiqIY?ynϧQ9GľGHe-sm*4yIxax ӗe٤B̴L0MTFFt<$j+h(uը]VR܉%mkE[|:8}E':u 58\Oߛ̍.b_%Ld&}C Rse}BPR3SnS_SRWhx9D+!tBpj?z"0nyhi؝M/ˇZ]=e2W& g,l9{HP'߹KWaBt(fl^DwMt5ٹEULk;wW;{O5PDsa3QTY))e9_G$"R3 &f;O~ )!3Y,H3Yzw1DabBjsZUc4z/ձZHVg[WP j6˼Iʸ()Ys FED8d-}8&j^olo UxR#țT iXfd0_mkTfͲ!5&Qޭ̖ϑ4b%Aϳ'=e.YsQ)'lcF2i2nm21k*S̖P A"dYiܱ3iF]&$R4nt?T<>VLcpt+a^It\APZ2ҁp(0/T*@V=RK1\5FZRΝ@8l4;Բ,ȸD!t71oDIa[ᢱaiWRΞW39LV N@UZB FV_i?^y~c$+]?MݜOXw˗=s"$;Z!!gZ5ݸN:s%gI)X L]At%+wA*3t-T=kY S י2?- ^`#zN,zyRw@猢Lb`ʢPEYe-^Nc6*L_EUWtkL[WhB6b8?tDԞ+$M!S`n~e^!.  2FV߻GMZյ}hW_ zC3WǕlˉnf?wFfhX4~^m`vq{o-tqܟW_CþpB1}JWv-cǔ'lm(FϽITҍ PlHJx~nw( XWa[~6Pbd %G߄xmΧvM4SF9HI%Nv3&$u#2yMXWҚlm|ehZV!'wIzP1b'{dBh^=LrNUpE" eT%D4:v(b(fhJ+MYV:S$v8] ƳQY d'u Z K '+g7a͛_yZ2|.Lf} 釷U|LMs}VzAe½Jc㗹t(F [<鮘NFA{RBYÖ&UKBzB@lTP\"Cxq$@s RD|C4*ҦytTr TߊMclDZ{W-oL'B̀W*:2`ʄڼ;GXD0ȼR*!1T꺹H5類Ż}(ͱ,;W q>+#׋_vHeDKs'weJ0^=3'қsT1[ oR͟ ɔQ|SY9sOD>k 45 ;ܗ96=̽;1 6#KA9n4cx'֔?COfgۍѢ.c!/#An^TA3먉+uev Rg;`F2JFNy$*~%֛w~o$W\Ī_ڒMw3 G`URHnW;R:tUkDv_ZiD*Ky-7V$~k"z\<|8x凩&f U1JKX`> [.8,[pH5NoeÕ$ ;XFJ;yCn Zw73~~\[$VFkbvwo2=M1aaDM y[+J,i0_Ʌ34Ռ ғiۃk zMΡ9׬dwh<eoќF U*ZVT-VǺےzgBJ˵_ ح5r1]i{ jeT<0a{Z &6@nc)hz?kqߦ eskԝUuDJi%6 a4}Tc*9/;X-DU&Jh4?ؓƬ^5&{ )@Q)w-{&H.8t٨a*!3Y, 7Mh9RR2<ː'yx UJBL*W\Cm뇝`^+"3ncT=-Ct#}ykޒ4L]bv.ԭ-KG$Uro aIE1ynPVe++jt^1,SgFRr+6"])%̮B\ F2ZŹq -Xl 9CgiW K 9F hV18E-cv5$}Fip0)q=F{#U`A: UĮBZ8\5Ujߵhj0(ʭ9; .mQ5 mJ춖\ifc˜W U,yV2P>4-XoUNJ[G8bNgX=԰\kC:/W0^ }ԳP)Vrm2dYk4vG*q^9j3^ܖUgF9Sx+R:nrx^j7z>!뾜lsxI,^~[yY bz$u:N'p:d'0զNn0'aи%B#Dѫ.u_O@+G(eա{jyn+'!n_B/gDֺa"PZGfoqX0q bI/IeP}i0 BLrTN>XLmvSmki~8URaّ lĐsq(9嵰Nj꟨^6-2!2&Pjٌ]t|iz>(J+^Ix=BXm7yEۧP2N^qA$Uz Z>Gp%+i 2-Ioܧuܭh1zy&6_'ˢ7GNzaҙL,R$ˤw1G!ޫYf)] ˺.";?.Z]?M2j5&as[?u]0?6S>f`o)uE4]1GOQԄus;=wz8f_#M9Oi:| K.DM/ҖHL봙]ioNB_~A+a_=Be&|b\ SdϏK6)#C7z%7|=ۤ?G5O&$IkVzJ)zvLڗEV.F`LbqԒ߮S*V%N$]v"i[FJWf8u ߲i.*si=I['~)n EWJυ%qN9Շh6LvirY"d*^?ԐXScT.yyW{=]83F #мX=L`r Peqoݞ3(TeAUhUe[(AD=gfA4\}b]iN\<6^рE6˖7$ ¡[V|Z ]5B`ejxi8S|F3b qMn\MGNm^X&Rd `Qth5 K6a-j2Z"]-CTM@&&Z4;Y5ZpQK p# P>-#i*ye]5<{3*p? e\(ew!SԊ\ƩO^yHjoYߡ*DhCH@7,ȁҨ¨JUoh1䭁U & RýVz}ެ+dByi @{2/M7d]K^AIFi.cc4z/(eSH!r f\׼ce: *wrgZɍ"l~ `"@[nr.w[cYc|~Ւev7[YeG"c=ɪ"o3x* t/mJLK86|yAw:v#{փo@Yr)R c^:j;g&Þ!Crlg͢" HzrがWvM8鋹cv]>YIi%ʆl&//F#h"ӠTJ+8!SA#:lVIN'6i=aZz`xQ5&̋>[eߴCr*e5JbaW=fǷ SG @KfDT{5J\hţ="|9`F9+~UMGa g M2l}TfGK%g;h0YJ ZTZ ZBB- W*qMRi#{K#g nbL4>%[JXD]HՕPP[ega Ng f̌]@sKvK0mQ8x^hqc[̝b6AufErPʏWs)N|u R`fD6.PkMO bNTM>g'Gח-!椗ETkm!{'|hI닍צgLZCoL(}E:l2)ZBJ佾"8V\wVQTEH~``졔v7;eNX"qc" U" 5\Z,NuqԍWI|D@|a1 ,JNHl]Ny .tY~1{1j Dt70ΔJxlg RT)5iBA%Y.qht>ld~7®;+ ЬB4h "cEy.žӐ{qDdƜ "dxgA|E l/kx|b Dz88)K%(Mռ(',X0eiOU.KƗf C/fa0l8 o0 IXQYejsyYi :ތyɁ] .]x5ƒ쮆ĪZ> k1y)OOm4 ~#R)sSf4Źw5ܚ 4wmjMDkaU|v]|F(fe@ QM}qq/f+ RWOF~s #e߾_&o|NKtV/Bn״m=-C"tRQ?JkZ:\ԡ>n2|z팿 ݘ=ZWc^`j/3ٽĶ\ i14ϋ3ݜǙ9c2sw)[-L wR0\0pneOGͺ#7ZRuJ *DIZ9^ bҚ0&W|FK>P5f TVn+6|y afNL2;*+!s9os1՜44ro_-2˄xлu0g3/KI8E=gtoc9$k؅;u9/&bjeQҬi|3N bӶaVͪN7S{%8i-4GFSYkYu(ҟ=ZoD üG]*bDÌa&QbMmU:Prx7 ҍ^7.Zo}$%2E׻pub`/xN/?ꂡgw/'2O~8÷ A b AĒ>)oI݆Rs`; AN]!2p~_΢*ʜ(3ڹ`O=ւ_IOH+<:eC`9%r~Xi[ TݱIMވeXe[Vτ ER(AIN 7,+@" 6P*^K?bd LeDsbb|5_~]IX[`%;)@rr;pyG'?Y)ތ| ;l}g!l`}<\ҎRg~ ,ogm6ڀ| ~Y,\BҕcsvZw&a21JC,+ +z~V#:'}NtxSĔv.2 < `%zSF l(M)) aJHU<=qb0ᮬ؇9hGg)SR y O4Bd-5J1Ў!'F9 J  aD¹*[*mi$6W(pn $brWS &Ja͖JR;MK_JiQfFh8oDcb8чV{>k''΃5)rfY1KɮeC`1Q@Z4FIWDSpW7I~Z&j3 ?n;|F\zM.8SV,`m QYivrr3uq[&CDo:m8n8SKnoA` đ_"eT 31ʼnu>7x` o~%t9!%9Τz:( l#?\|@\=ɲ)w?oz XbR-#d۲lg} O `(p3'\"OvզZ:-* w91-  47nؠ%R{#a`YciCE[ bTYYF)s)mgs7Zmx1p-0[t&Do#4N<(A@iIbY=tjZ|}6auzl@l-` aLƨÙĀh-^ Z ֭-ܜqORLߞ*:1עAEr!X_hsFI3'% `K2 $`3h1Ǫ\.Ze[>CԣϳMU%m.| 2-4ӬlC>Lύip.TSYF?((D5R Losj\| MpPV<*Ő%7J\lfe7+)%&Co~K]cl4>1"7 Rf@LI~Q !r}{QHyz<%Ka>'y)`ZD> 78>S =[/F^ԏ`]g:1e|2Lbn:Uf '\ ,{4F&zSήC^:C+`)},^%yU3G zfNg(#gvg{+wxF?>+.5w4d] G&,xSyL: nt 9}]dwz'P@t6Zk(Ss`nrzNF,kM2KFb,l9xvaTdGK$Sd!^l0˜#W?͙RAiuhb< $||5@btzJ;>#> DbgTi1J2>02QZQ(֌U}*P^M'|g^kŪJUz<4sy&hˍuyg ҨWO˟g|TS}g F ƢjY+VPDj][o8+<;l`vgؙ~>ЅJ_Zv;^?E"Dő5 ÒŏŪ0A,FmKJ!e.їJ: +e{K.=V90NV>&Nnq_\;| ⷢiQph lÄ`ogr0w?׃8ANwDZ-A[n\2! eRDŽ\$]gyEtPpۓY&< nh>$D9\&'wQU$Bç%HH*%ǁ\L,Vj}e> R|p 3q";o?M>Op-qpY0L}Q" Yb\3 JROq桾a54C1)iDK*G@'2Q/3 `:hzv/EtYecs a uTD ʊ Q[m00t9Yn 0x c*3p~m<Ϟ̼(Д=شJL[)Â#1b(,se8 U)ZuGJtO89}I;Qg[uge.[Ԑ&D@ a8R1!1&a՜D(:O6^{wL#4~CULqp"86@u4bmjftŹ9G 9p9q3DUҐg!* Ϫ LR`RJ*J;NvmS4Ab֥xPkNb4I1iQfQhfO9OE|ⵥjO$?'ӽ;iJv&5?+-Ӕ)l,04Fj2#a7p te/?窄GzP3`R34pY\XJf(%1FE, "2b`W9mMԱ֒kteR&+ΝFx]_aV;/5sf84f@tiX ISIjO6~gldnDsmLl[Xo[=lFCC3 {3Ո*lHFYQ0:cF9I dG;n#6Og"-F= zw0p$;I$%CRRUD{ ó.7$[_+:|_\<_jb͓ R{Gg4+razWɶ֖Ixr_~L5c,NDs.{F(s~/5:ġL(8.4]J` 1Uؼ0άLp|H /J;clzM|KۛP(RRCځt,޽+e]R[0ScsH[Űː,`]d}i-iyI04F"Yyu~3tY uRKk~e%sjHҾ./9;rGh4!׻%ah` l  &$'a t݉Hp \շXɂ3XndvQ[R3v0t8aox9Yy Y =I3V:]_%]ChM{MӸCHMcq~m"s؋ Z,іs*c x\x qV6nQvgeG-;1YyuoUaքDXʞBf2fqTAY0Sgҡ h=,Ij$JRa'+V-Sa63{n"])˧-NGkg\iùlAbCK=v)^U_yU5EHG5Hiܞbi=,kL; VxʲBu)D?(u񾪥n^b2{6m4GٖO[>+s*RM$E%QSA%P~%\,e^J8ňlHN DŽ \/sgsf#].lz.ЩCzZcF#l/j%^;JcoR[k@辧pzϞf޳Jvcq벍7џLv"g7V {P٪~;+~㬆ί}~*v*eMGIa"8/Z3S/YT`Z!mSKEޥߧ.mlha_"؀GO/[+A`I? |QoG1!-ϧǠ)~dyq\ؙ>I>G6a1?1V =?6M)&ZxO}q1H߂ LGp˱Yz0p8L`Lzy7Tj%U<"Ȕ8Bd+becP {zh}ZڤK3N9bٲ\kb¢sOhwSZ='z6:-_[_c\ø4 !Jw?I(XR[Vr6zRnvH.q} V1b)t䫰1,|EJc~Kjd;G1K~h諂 }AyYѽd3OFCn.D7aM0!FHr qjA@L; Q:9 .Ao.No)@.%6RӪ0z9?S@Q'8( Ҁ.8 ;<~_Ht"L"s׬3 9pZb3N !O#BTaK+Y*R1X~e|Mĉh.S]uQ*yl9F?Pw|I-jhFVظ!H= xǩ*<-YJآM g:B{U7\&2>zw>j`pN j;f(Wk3x A aAlFP?@{;?E?}1͂,il :$Hܭ6|:|qҐ0<|vLB/ϼ9;=a'v\r'(vYțҽ幥$4y4y=1xi$ _Ol`Oy4IK5JI0܇wҜ+EYѹ`:1|/p(|9سMQ>y2~ 6G%GڦPIe""vÎr<& {t6$4 5(%4%ÑN;o?M>Oޙˬoe4>\!-og8k+o\v&,(nm5\z)S}Se(y)] }ceSxvC{yt W,>!v@}-:cJ F_B뽳x/}țTUd纎/%q: #%i<˄U8tep\h 7tfMUjʤZ<9$y0Aϊ<]*WGw,5H~vBBڛ\G #E*6ט~ =n [np}J3r)[QC87f<[Fcg\RݭVSNZ}iNާFᕈ`9+ϋ`,ɵiY6w6z~Z)\GÕݺ 1?*b,fOQRN p՟v3؊ E,Zd\ i;ȧ\s/} ?sbx_W]DI{> MvCYx84WJq(gRn7YSIU rAe-' sp谟COQ \CO܀(3D+l{\TTC;Ń*ldlH9F0ϥzgU M㞣>IOkD^KHy# U]J$ővgjU9UY h蔏힠[iqڰ kKʮu -uɇ[Æ{ lҢl:އp.uoGYaPk02Cȋ1ᘊ1ߺOa?UׇANK`to~+hضkQ=$am -c(8;팴퍆PV/B2C;"2Ơ| *Av=:yq:Z'"{ mzǩZ,+vW:dm+| NfVjz\fnM֗jA)OСh?Yx:SdU_՝O9<,z03MO_R)}ў֝z$D4ΒHhR7'-Ufmx@A2%mh-u(#/ѤwSf ^^{?4u3Z#gO"P0X5-R6 xi?/jru\V?#6[.-a.x(y]g 1.%|D,5Qm r=ף|MZ2(s|N!o& $Q-d*qd\~H xŜrcwM,m=:g'Xa e M{}H54[ zUz_t4j&xƨ0IPG~\ӿ+Yt>0-I*Bc@ }E&67,bZ7B}T%x *x,,pQ C*]ZQ)SYW@Uzh'uJ3e)RWsˢ 30H2&l(,4W#"Y^UQ *>SF_Q {TVEAAeK n9@ ޖ%v* KQD( -$Fk%W"L& 1~?D D[<O&tǁըu7{ī FxѭtybS@rA@`$h'˄5Z/ f /LjU+S.$8YuIHbA_}(fNmJ\ץݜ%nd䨧7UdIJ ""ŌlKm1MN[67=5Dٳ8aU;]+\Wgv"P1~pt@_v^RؕL"j93X,g:+˝ JwPb]_mhRf΄Tٕ4QlկSCC!Bh;7=* Dv FX"VPH|FU^:%(Dъ-)dBC=C!EEhJ+mДzZ5܅-g-x2Ha['Z:83tͨ3.DxEsf][!ori6\Zf1Rtjcnߍ'$ȶ]ucv-7z04Ay2z4|'9}WK]⯾V*Ji׀+n{|LJwȷzEL|7*G2*"jyZ LJ>,3E|9ԯctKo*f9JP:EzhI*]Y'kY!|nRj9i*5\yJlO\&}F-lh6E[Wk(EݬrSݮ/5mt%"8ay@Et_t;yHmYu@9}Mʫ%fsU&}5N7Szlӱ4'-]BӿE,}^4t(żi Ǣ>l~[ ޸&IX;T9R\$ WG[UWJf.ݔWkaRt4JHk֍kޯ۠R#?Ř+$cN~VF{#wnnR n^+o+, Rad(əɽf;N&$xʨ#und?Ifˀzʰ f Yr:MSj2DRn/= V xDQ E|$R<ӊңsEC*(N|=n Hi >OY83x~Ģ+Kv1;[#Y%Y=V ?BxCuR#mοNǤdÂ>Y|3{Itи(Β#4yAVim{v"(=}^T:V&5Xr;> 2`+r%eE0 0ɕ#)3v&<6LԳ:TǨLC-N:Ϣw3g 0L]I\e"TEͪBz 'N+Y]q;c;U6ΠAi3k}Oc1s+)-;Gڍ@CKE)9ϴ,3g  ӌ?WS:e54[z2 Γ1\(" KV記2DJ ziVMƛ9 ;M ;J sqx8Z?[б=nZ.x{9eS"Ee wm,XGh]%{f r#"$:I <3Nk;:*8̀ihXt9=KSû:a&m:xYw<+g`[Q[|rA˵:9CggH=jśUZVoz(φ\_o AZLӨ ZZ#D(r-Ϩ&ೃb˫dZJx`DkVTELZ% Ewcm1*P;d>t;.k@V3̊Is:718fvU4Z-Hs-vVlʰy!ǓYgfGkkQ:p:~ "LR_]Jϟ9\wAtƃ!מ*Bt *%Hn# ]HGۑ>΢Qh nz508'L:T8?,Lȫ~t,XxVjX&x4#g*L ծ63mKiwR[ҵ:pGɎ!~ҋÝeyJAuwKa*#I YSd1^J$d*ēh\jm/15 s!uqR5kѓZhf Gs3UqɎDlZd'`rVB lrϏtH0n|3~UBZ8mbW8ORSv3iJ|oVz{r&,IYeb%z]_f>k[+3uD YM,Hu*࿽piպ|o~ؐϿLadP頴,NL4b_԰/j$߄Ub~*zp }|%m@CykںczO% 3f XB[ɶ[:=?<Po2ZMj4,Hn=g;XV|,SOq%E%4;hn7c0P<:RPLZ"B.c,6ٻq,W<,Y@?,н桱3~X4I'qT=t|m٦,NUThtU*xnʭd, 0|ښJb5d݂(Ŋ5fƁ{%5j4Vi4V @J;3B[wۙ71^I*}?xAF1fL:c$JC)!>*`*:zdOb<kփ/̽g}%80&ěfw>YFkҺOXΗL8\/.h$gS[Ӈ*~D1*QRӖBb`Ihp X<؂Ųax oe,7Іw{6U^,#^Qq{F4"r/bQ:ja?פ*8 wY>N_ŧ5g|cvdΌ׏o?g|5t}>G5#S\|yyQN&?_v;w:J9YZqR noBo:eVjň, or4V!y6WFpD9$#fnDygH^ɫ %AH=" _G[Whχ %z6Yg=%{eog4l# 9\Pݨɥt`hѡ+ʭD^cq/*m@XဠW'+:-7:f&?cĞCaFfzXRVt3 9᜹*x~9O$[g1"dVX#xo璷<>R癪4 YR@QzDCF7=+U%"%{#Դ#nG dW$u 6Ouw-69^I͓+0x!KgI{RK1d7'W uCzR>ɵI+GRq*w~EuRUJסJ^4@1Lʼr 3mJ兒30,<pZC2Wh6 > Nsn pnjxWZ>dIAX },Mkq6Otk~p7>=TCmM8dY{ȲҞ9!Cn YnMil(rݡ)\oCWsYQuU7^Zz}V kZ xԥ2!'ڇ{ -H Z_OAhAh2E+}d-#јz居N$*a^4+T ]AsZᛶb(ze_pʾ\`F3N2C+j}Zh_N7P]:,dIŢ'n:5]xěqyQ +OR1:8s&ɶ\2vx_S'BRw.rwF{ۘpi qMv1945X e0j;sQmltRH#Hl:N-F@Рt]a8^IKd-F4֪){3֙ op⦜GSQtءB[0RӰE Fy!>>hf%ӀHWzbiG.'DzzE=gʖosJCb=hghV4״̷vD#bjHT &xF'oyR[tqcJ36qcqks{C|3}Sk*ጽFUj]E\kl-PeVV*̹ǬU&D9hWV qU0+$nlUP]eCa0.ey9rZH̢Hڏcu!"6d;Yi&=;w;~SJ]]vU:]7ChUec~9V306v2NG[ِuC5/g?V16ntx o]:k9\Q}mJm5-L;~mA:2ru[7&׾*̨"W%387"JJrx(bUP'ostν.5Du3Ke8(eL"\[ R;7JHA^y^#([Uw,G#7:9pd^.=~rg% 9l1"]/'x!͹Y0 p#D9;94g3t}HWHצ51e8ڜLh)hiʠ  R 9`>B[IYFq%LY-T dc{UT7FßCwɣF]Rxu) 2D1]^p,B9!cqVW-;ZWΓzšbU[[tR2<؛<^}Qy;d|w yTm n&c><6)ofĈUNg 30Uƥ]Aqf %Y9$B-i^n#]W ܚ&ԝ.}ugcEgu%=?*V \pvZGw7V9iiZ@kk<8O~80 MRH]*PH+rRC]|CTC|;D;; 4]j@W'N+I.+b1Յpa.YgC[Ihlrap|F|s} tTэd:+(@zſє_./f|9)3B :.{~͒wzNoo; b,R+2{俌GF Lkhc:%ܲ'{$lJ<.#ӵ22 $oXqiX+9ZzK,.\8Aex}4;Z;~s5ms݃?爹f|dק%La"J7XA/bI~8uva I j;i5oS4=75=#Vڵiu7ȯ")J1thArŸ:OHzi(CBKYPfw!J9<[~ yB4V&&&Ʒv Z8E-s`RU쇆~]+ݭGa(Y3Ld6Z05Mbb$]f=FJEUJW9bYaQx!rDzvܡO+.(eM`A`ԌK`I)(6)h+d]E8%1kVFZU'SM"k%+.xSd" QڄזIvWcTFwQ?{j;;w^aF~,IyS7yS7F]$m@,RL)f &gKO"FpF2лV.ONTdBn<ЛSFs@cl@x,{{n2 QYCI*ڇ/Z))6S_-5{m6E!B-WﰹLގ;;3E*((2EvsZ.ل2OAJ{ }!E@b<6j.CFv6/tccEdcPz9WqຓL{{ptۯ;K=xEo&DaȶϱS&:bSJ)wsnᵓT(Ǽ"Fʔ?cf]$0Y ׄԏ(T%>YK9+Oޱ#,ACg<4Eq)Ii[%1 :R5pkŨeRx̮XI'deXU%-OW5ʤ5z  (SU=S#ݚhB5ZN}hens<gCr&QRP`{{$xL-/crJpje&O X74Y4K% ʰ#&At?ѪKk?֍{jY*vÎ辇AsZDl5[:cV b{!{ވ5\[VE{8ͺi,j7wY5w=Hp9;lJ]‡Wx|{:MMuU렟UTo%MV`)L.*tlm&x6u}kq9÷WWc)4G i o]C3?:7GoszsXyc~Z6'ky|ofqA*@@I6]iAž]3ZwJopL1lx">fH.ՖK-iwy&ȶZshK0,  0k9g*: ȷ(#m4,_vZq|zu髧uC4+0u`T&na`cjkYk0t^Jc|4eZm_*x^wwK W&1/^.`/y'{:1\_DK w<'`cΤ =GIߔ\Q;@HѤ&ܣhb'o3 [Ӂ,<pHYX4ҭn#!Kda k)YZj|dA- au?Da @_݋;܀~?QxUYX\F9DQe\gz-˝zޗY3'`:Y3h2b}ۄK0?h(xj:|PETAȼHȠ@$1ds(#,LYGz͢>(gS?iYf#58uk#Ӣjt?ۃOqQ^y_VQ#/7Y/kW'Mu۸4_-/WߡV+1ӑ(5Pν|Th,>'*{/f GF ;Pjm> B2׿lSZa~HhbD*/ӛ~><9,ĜS5eRy|X_ӭ?;<Xʆ3ǀgD)fS,a3J`pI;[=loh_Si_Si_Si_Ӧ}]dۗ(- h<pꇍa.Ȳ<5Rvl/qbkt<мh f2݆>/kjws:']mtE&b+S֘Ue W\їRH 6YIܔno)%JJRBgab_ Beiv7!ݭ7Z#&csY~K}:Z7خ37\ߜw4j4$4Z:X)ZCvP5Sb~僈+ۙm L! ynb?%QVIoPƌ}J >gu %~,ɮvf4C!߬CL;zV{mGu5 b:CLst쎸63F*1uTK)ʨX-*xmN&hfbEۨ%{III\.H*rҩ mD{JI5tH"DŨ˕$dK;U̍/Ȏ UJ֏)MJtHb :/jDl~Vq @)9mK0H6RM]. SnO:LF;NHABE)s]h IWͶLm \V7+׶X,tr6|ԵX#%͑8uBZm=N0yxW0ڽŪy ^#5Yg+6J%mhM+>޽ǞԲF6H[odcfJO?ծSˈq$#? TEF`50H!7Pg0)J ҺF굑÷_]miؽcY3x.vķ"'x>k?EQr6u< c<;I>zPʏ[F҆g05ޗS iID ю%b-%Oz, DQ.oD79qψ ۫b}X4>wNnN,]xۿaUwGjkR%T_͛O$7ue( &,BݭԶpsq:n{쵥%+;k%(uJȵĬ7{Ew}=:-mYb|`B[";KdgھqąAP'w7h%1G F>y:=C}e"Ns^. b 7Dw. F~8 VHV8xP釯 qV\)ϧycHwZfʨɱ\yyysAeT#T4Yhp юK rP-etvvr8?oQ"}xΫ+& FDu3e[C t([gDRT*@_'΂eKl M7 yLQSLmHM<^8'nɋ.n}#Mx;xFya؍2i8}E'hCv:JhmF+0u\[&n=vh s~zta˶r #/@Q;givz̍vEG 1[ɕh8Ρg"xyȏ @g+K+suOd-( gï>;b'_/y2NyݣN.9qO pb'XZ:pō2\?A9W uXä}#7ߣ$x-uЪN{ō 'Ї|DxSx)Ʃ{"#=o3vVL$b,??TK[VNE̻7|2k(b뗦.QZonvozqúy:HIbE}05c"idUf-f$.zN5QI-)'WC$G%$hv} WKwNa6_ ȧ9ft6 K[ţ舝.x=PAoWNÃw 妴ƷJ1તta+$KuNkl≪9.f/6-QC*jǔ|м̇^;H/8Ӌ=~:(l35P+._&EY#,umI3լ9 GHrMmXo,Nϖ^5XPVDI%#5 70ԠJ8 d5)=);)Cm;E"N}L@b4Ie f>2o@ *TCO#e *=RQmFT9~6K>*xߢLUWUD$*B'5H` 4ʡBB9TA=U+[IɊ%rmcs .cC4?V1*7?vp ^v≜xjS+#8n5^JqFswp{=BuKCһFQW,Rz#&Q34?X sLű8m S{6c߳5Bfs>0eN0E3xv+$T(Wac6As)9´WB!X]rcz*>&`]$+IAJ&]*eN(ɩqTĜCp5DMj;r22&ݰxgYr^>X$ Z5 n8׋+.L:pr~uQp|@ͼ&oyj^\|'5TūvXU:5tyξ?vr3qoW-Z҃3=,:uCZbv%pLYSdG,< 4u Lީ{qP)j/~BN/.s{!޿\q inS bѹp(CT4}8ZWzZM:<|iSSg킵Ӂ 1!>Dn"="b񦁸K1NܓhwޏV R~`֯~K>Ҡ< ]M5hoAb'oaŰY"w>XN%2U~J=,gE<s3u8yl^fr"sٙAwP^dr!hxg뢗zKA^OyZO3/8{q.Qk6;I%N#@Ff]\t4KD &I-PuESJA V`G6J R)ySޭF`ՎEtP/<5-uc!R&K9NghۡKa_:Z¯K@0;-% 6zKu N7p6\d{9vvҾ^dbi(7n{7_x^=(¿4j^η^YlDI>_h!o\xZ"[֣UUk#ޘ'~=%0G祀c4yy>Z4N#Kjݗmfᆅe3Ιw9˸>j0ЅN7aAv4ft~(Ncm9Ebf; -nvMVۧۘI$xC}eI<$8@&Rs/y`G0/s;GiG1[5GםgpFo[\8{czS?_PuǮ>R⮧7.f-d& 5 XVi^(347!Kv)iAJZЬ7Qϼ@m}]͈PmHf44`(sW{ǒ*F]mρ[΂B{Nyͫ-X_Bbhs^& |I+2EVXzHu&5gGѦ^G 99 ZKE*L1Qңt)oZY2ERjX{qXḳS뮡.εji.Va gJ\f\fVV3]*ȲaCŵ7@(#D6EXp8{52 _h5˪&BE˝99UA!e;`ȶ ppXI'`"ߐjwͤKCw&7_5Z n* 2̒2FE" US1A)Ja@eW]d X* <7^4 jн)hY98ߴ1fw.A4iTvt7rbZmgF! tXaVh*x?хi#~\NO\:9P>럜8jMO[%cRG:M,4(j#4 0as  y[b "ܸ#?E&aH0io^>+9Z2g/6ݥQf ͨYaKpl"׿LOWcޚ/F3*gTϨ4YǨK!8eejl/ՙ F Z^ (R 5l>0e9UK_XRs}UyWX]׀ 6e+>{%b/ڿZ*$| cilg;[}E֒<6g]Օ:o0::!3w ԸNZfC76\-^:?4?]-.s?tjl&"\|<ߛ)64 [U~# @F k V erygZ/MM?$)bņEkOsd͸QM>?mx捳oͿ~z.͟[4 ݵu6ۍt]p\W[Y[6{{md4U쬑uad>:JIt? ePO[kCt\( #Ϩ4O-} 0`k* Bz aQ PG]Phzv%n"HԮ+k;kGeҒ< m}|^6AX!i SI)6E-as?b#PS6W14j>1(~L DAeY4{17wwV|}ھ7 MWF^G,#r ѻc-؞mtģ$1ssP+uG T&v :1t~v }[o8^߬Qլܔ/x-'kQY% [!I0LPJ+z]'ѯym<{ ^_LNpg 3urAF~"OHI?rѽӤ OiAH<)qӤӤCQb{Gz՛β۝^/y6ò!oIK)Cǹl?fRMl0cԤ"Fgc'nq۰ܪ d5AINЕ'^Wص%I%>? 1gE5~VҴiВ|!xo%Sl ~yђZ?|N^q&GY?<|8n̆EDZ!h]tu3iI%mG_&A)u.5;n̘C^j?'Avϵ.ֹ6gTvyT}B:3[aS\*#6{j8+`99WFnS:E^̔9Wv{$cnCubVs8z(1O7]D4nŷ%%&pn g?N}Q3ROu'D†?'B|ͦ%6L_{gRAGmυo%$i v `v0d\ =%kwq^^;duO\i-}{yUe*ߙWf~`pód<$57 qF.w~E3#h2L *pIvuFTeӽQ ;:@g QT Y[>^$uƿ5g&@O$Y!L*nLH>7:ѳ"5r*" ܓuY%'j"߈{ws$̕y|PBVbM`W6%Ֆ qh pOa6Z- M|#~pBhqHY/r/ amo$k.`9򢧁 03<-g @bgw\~ڜ-ؙSSQ1Xr&5Oݼvn^.;u냆᪗<&Dg8yh7?P`Ch,`,6GbW?>N)K;2_}YL$<!T|&1̶؆d1U+-C2Fv8r[J6@uf'-Ӎ*"lCz6F'f}%hX JyZ#Xcbo_WHdx>LѻZxr!򅡌256ͥaDո'DNa= ^פgGH#a7Ĉ)H()V<DәrK3!58KtULWL!΄">0ݢ})(Oon?hջI6HHD)Wi?YGP,9g]CB9\b{)Nѷ%jhe%[+ӣl5q [ܱ3e̷-qbύ|&0M`k !H8ԎMʓA1~*M\0(1ٻ޶n%W|]gH%@QMp@{Eqq^'N-peY>dIr,A62M33*J(JMBkqZwRy,DS"f0@PZulCJ Ry뮊DG4&%3Pt+E\Rm;\\wpۨXC3=fyY`s2>-UzwD;8_!TH:EHB E99^îNղp<ΖA\Nw,b$%l;]tl1b: {x rOu!ukR>Վ.K,,  [ج˱0Uh Ow9g4GXuv(lLVmdN5Y!OB jml[3A.G$A-!ރLɵ\h>eڌAҳ}3%WZ3=8pzd6y1x 0Ƕ㵙 is#w՛յZ/:-4+B/#|h,dfBztcMQwxհQau+j82 !!#^67[AXEuD)䊲%(V:21T@ E,IG9hَwtMiCMbümrpmd=ww\lA8a=O8Qo5y96Ʀs'є'& rOv.m{'ؐMz Ԭ˱) 4 -t_˦6u'ȁ \.hcSA* 2XDWZشɋ0ûɋ Pus 䀅'0TBg{&J9p8Lzluh3>~L\x:>-uIG:DqyǬɋI ^ih#L$M7י T`i$Ļmz6!R+m(1MI%ܦf\,JfAfdM } sXE/@Q OVXD5 WY״˱.1A0*Gas&2cV 1^<.KsqAaSYcrn,*s5Zܝ &SF=gsoz]$&` k&8f/3e .}?` !yq*ȇC>OA[>`[$b3vqeh7qs"h9s~NRqv-7) C=v !HDKgD42;=s.lyO\4%%:(3!A[tݳvEQ3'B}'eh96bX+]}*R?=̥7GNSVro2m1^mI=*e/7E⺛aye>u=k֗u*<ӢnsS͵}ZȬ8(;Pܜpeja]9bQl,+n%4Mi=.*虄G.T )Jpf(GSbўG{zS*v=/hh bX=C LJ}7'HNwv-pyuZ[cXf.w<"CuE5RTk*uN.;["AZ'yyZhlq6UUpl"KF7ɐ9${+nyvל!Tj  tZR_t-XVZUYS%m /DlKvy%M5[@I%@ $gӂs7)Մ_m(r qߚ愪ۮT@(:Bg`bb foʭvQXTZ>)t@HeׅRQB V;x}WcؤQUs9kږlV1ڦ uX6 8}jž fl{u 5ڲc@FIQOZi=!3* aM6H(!N k4sUU;ӯ`ى;iKi7٢MT jzcgvԶ mcD yh`U"lUi:]4}-_jTǶʶ[iҔmI])H"C (UW.G ֩lVk%P+^ÑŲ<яb1} +-j > YHS -3F/ 3nU#̋O=z Į*~x'~'v"AoJ+6zB9k.:f2D{QK- "P{Ѧh\ۏ"V:SɔbKI{M^D'9wr!)7,=AI3D?Ɗa]Z&9J i%Hoby28Ozw .ܖn9`KaK3]>hg3I Ђs`S.C!&L;35FCvfl[EyV7CO);8sLz_~ld*Gŧ#ؑ;Oν'>J')Un/OT=*t꬝Q4Fe2 }&!wΘP@Gg( H7uz}xC,&y!g@zpz/9í:gm=j~Qs"20z0z:-%m)ѤB-!DS9p}SzefdCf^3Tohz+ xe"25 @o |1'.9|^&!>)ɏ'u*-9|~&ᳩ}66B@2Wn[HաaWv^9k;O$=iWOHke!8=["*5 M ]*@n Ʋi1tj_Mtض;,a]-y^Md&HioH{MHyʹx8=ѓ,"8zdFqKu!W# +0L[jzH?c~R<'h>ؠCCZM* NrhC%jAQ{ׄ/>t`3N 4x39\ K>L)O)D:P3;ќsv^]?S|W2RCb'J>fA2 ̀tc%3v |8I]\-,ݚqwvX0;35&vf J <0&ydhR'*.dh\#@vsn.x ~?>@ p9C nϡ0!A;yJAG#f V>U= /3ussnM*n #\Fv 7>y";[tǶa!;\!FeEPH;sk>wۏ"]C2RTFgaz'jofhw@n935;d3A!G( dhUa&C "wŜcQS9FghĿT}v吠ی]}G0n (dd :Fod;ȬNA-?N+tMWyO*pI NcGL*8"!M!oȿEL?j %G{#w=HUm"%8x[HBhTW>G.Q9dĺfdzPᨛN{5:*bMh{_cզ" ^~=rUg1 ccRVλFiъJ9\>+bm!cD{"G=*ꋳNb 1k,eZ]$z~{gy-o/XHK糯'Ň_sPE+\(ЩeE7a2֡6ul|*APF*=_erLǤ~FTݰYxןOo/^6dW?7OY:Ķ6/W,c/`?=F{Րę ŕ%vTG'4>ڛMlcʴr[&lqӼB5mm *.WCo.|ݝxyOBFvzmq#+^ݼZ1?`'Ha&/,h9Cq`GG&ڃ(&c"``MD*=CPE#櫬ծ/j͖ɿ&ZBʼɯ?>]}5|6<?~==BG}SE˻2Dn?|qD Oyt# <] t~r6b o7׎0Ulf}o=;NTkGhWIǍ5eªVltRg 3I}/N.M bNT~U<ؔm!uEۦt̝bʿs8iqk*][vm}\%.6BDH޴5.mERT*K͞,ԚSҫEi m%Ҕ*!=ndDDb.|@xP<(\t1ekbx?LS/D;x"piDj&r7u(ƨa%b&o;mhus:N7 Jj8Bٳ ڕ8)squ;[Dž]rsVdHYX5{zOt&Mz>pMc}$ " B3c|`$7't|Tgm5(ݴSSs!4mJp~^q&+FsiPC6Lyd'T?.,u,FP5][_7cy&˥ +.z-ù뛬 kڼGdT]fէa":52n(Rg|.RY>_80Z4u`x!>+x`ZgûE1|lpr ӰK{jc60~G{7n&.}p1ݯp܎n"JgCL۽X:tP]Tt+L{fz"I@ZYeXu'q?l]}oye]WP͗<;h2zUȊX-(f/m2Q%oa4%SRJt F#bUUix"ɘ6'ftʻ3:?t~N}I̗;V}Io~wd!fMC+~|7Tɟ) 1;)tjGke0K]2`e`>](6 9l™g莝ˆjzNķ?]>ǰ\P\DҊ}7@a/xg{:|Qn0+ޭiߺ+{E4"VOY!|j5uԇ[}9Bq_t_Sb@b!M}  aB}:FcA8"% E I YȞ XDF Fr,TH1+}ma0ʆƵ$:D!WDP2 Pι'ٍ,],Oa/b݅2wDDi7]7M$-6w$!W:;2w`V_D&RH]k%$֮vgz0>[o|:X9EgqZ#)a0JFrSWp1#h_Հ,˯C#焺bҁgF$Pi*JOɱ?QJҺ#dw`ځۘ1 [ӵ}Hm 1Cj@u_'X-?-%( dO= Q4LSn;(N&ٌC89"X?uS_icBZbRQ0XBH Sjk {{3Y-i .ro. UU=qRm)N̞07o/:,RnI&mq缻v{tOX뻻.rV;?3'ӟgLόfD{qܘی)͐ҙFj)Tx :nNG (U}Y\8tG~i?߾ 8nݾ~N@nu.?ooԧONN -B[Kw7 #$5kǝNGLJGJ)-!&8`avȅ +0@ +a$ ]ӫ(>vV"yElʙHH1I %}#n\`Q6NP.ar[EDnI\uWUL\O*7NP}B4e!?‘kNၱ#n*1PpZa ""Q8͑^g MZq?ʝw愥L&y}Wy;I/ P8DX,śK0w(F ůoobF97FfۅagFSPZ3* :`zxA9oҁ!Rˤ (gtT5ߛY3L,ɼ֬uǛ)]p:1źd2X;h :w0;*=HczFA6³\bnz0%1Մ+✑K00>\@N5p; "@2<ĝ\ƌt>K 9oEFx(:P\,C՞hfsB?x̸rиFPR. UqFV֏[W\gjTw:N0nd/~GwdsXc{N_^%tc,ɝĿ_AmLa*'3DZ;i^bksWe;u_$ކĄ{V1,L 8䙳h3ξ޹g>0)y_@Px*!Jo )J,yΐcAR>C;3!ep3g<jԭ߁~ %ݷ#3R̻ "<|)3U}M@?Mon,kFDB ($=&BSF+ ->G(mέ8rArblBh)-N~D1|kEF`с E?%\P Y V*S-/~dÇxq l8u#!{|XrŏqS XKTo\Z\${|rF{;2" c\aq]y}+ҺIH)tSWpI1*k:猎f <ɰֆrpIf.:P!.:l 0 Drjy15ZxʻOM^oֽ6f:N޾,=לܺSgca1qx9!It7͞eĸ'O;=x&ooy<#p%(saG\q2Jz^ہN;x?{: I%S_[0u(4M>ЩkiI.2Xf̸ % ^a0|OI=)Eh5|nbI>{;!"`n^9vD#Jvl+w&~z `s8<=zxؘfKۻ^c-27+ e΢ ,Nv@ލYs g:/vi2_ o>*#Iec^HUce H R8߭fG} `CP಄%^uCdyi8_xY 8v0UTX+)n/c/,-?M\yL? "3f>]``ڣhq)(T.⶷/<k]~*-="Y,U֧vuuvt.CiP/BD,_=&Dc&~*fVEgV8v֕u]ocR`]Z8Tm} tkt #)cZL2_T8/#q?zxOIqbSf7sŰl_ SG |j]I?qu+Y#VʄO?xs` /xR }akYeT^cqvL{#>3$'Y]13DvovfP818, ;'A$ #)PftO:˦l"|Nfz(q 9(2{*uiD.c)1{qA)Wɥ6g1zYgj ;^^>B%/BBz盐z2> "*$#[j8z_ ۿ-nz&ʲs6&QEn2Ѳ9>}pqVl !xU([8P`ڠ`ܐ]F6_%6Ghc .mkO{ T hőj@T;" hAxҊFsde_ &@BZCiLAjyhI #*\`ВF}glY ?SChל  MX0(A@Íwctoyq-\M EYm~%4 bf/.#&IQ ӵ"wwy cQpݧ Xb H1ybJI50A0kF vToʼσ`< GZT(R J͈ٝrDLw2nǐ!͍MݕNJwFF8>v|?9\uP5TJWW6 ^97Zy&u(2}[^|\.}",/.aT"ЂE>?v2-p0pvS#_}{x_~P.S) 0%;{=5#7 Q -J }VImy>K[Y;u[֢E&1B?~g.!I@Zb3N dS^Zv9#'PYӛbi3[΍Hׯ֣t~09lQo_/7, f狫8_4KG[t<8Ð;Ne򽭇T,p17_, b6Vn]ʲ2 ҉lEZQ>jܷdu>]=V3gwIC>7O 4Zbqo}mH=FQ",{JaAÿI+^{הG@"!wz-3]"KYEBJJ9t4LhEj_=w͌ `ԯ3\ i>KZ̘k)^g_q_^@BTR E r*TwqUcrT¤Ē<ަ݈.Yg5U=?ӓ7wiZj|oXTٔ,H{F RV3 }W53p.^LA G8K,wO8Z?\M.D/Gd~ܓ%-x܂J>|Z>,q_C?EMU~G-䈟]0X>?56fϹʕvP"Y3j#'cq-n2SSS;,)іۄ*AqUvp%m6\5Ӄږㆫ&3HPeЀ1F2q!$r;K)Ih^+S[sor.3(beh3Hpr"G!CL=E0e`J'Mq2ca?ۭ LGg3gOq*Vo8MI+z>t/*fnW۹Z6>@w&wG.6dِieC@ z,~wk?4~!K2"~YOfUrpV L 0=R7{t)2rOc$3!FJS I0ThOZXM[U:"aQs7 9B["+>r "$9c(oGUϬꡕw,ST `% 0?_4њe/&:'yq?Y Jꂒ..'Nϗ}8R䅝Y-Oib*(;8|S FەR\j_=+.R.YZ%K9؉wg2TBv- X)!Z+LqF_|i[Q:y6=-h)-^Yr:R!n]U3ԕL*N,g16:cNI->TcBw,@@+&edoB i<s[@-؃Oh5DE4ej!ڑ8&f{x^<є0'T!}{VLƳ\ , BN%Q(.kUZ-HtȞ(9T-VL+]ů%&Y˜jSyG=.5Y/-h0Ӡ C2\1rR#O# ],g$ЧXqVȇӃ/!7I{I&0(ܸۛޠ%ٖbڒnb ز$3 9 qM6{"h#pA%>d8l`pl4đQ%ZrXNfx)p!KH88,nZXiC`\ `eT1r z`wm %DVn1hIjISIK\U,63SƼe\'WqUaULk_eԨo8<Q<)-Xl|X |Bg0& mg+lw&±\kqCib46vSlKKӴӼRxѳjlG a `ϦGi޳|ɧA*<Ȟ_ r`<PDPّuD$D<8tPRv3%e vZSR*`Li'IK8 r˳agZ"g3m,>c2:KʺxT Bx pdc x݀ x*~U:2nxf G `\Q.uꥰ XԊGEKG݆jIGƗ#B8O(n<**u4IIIIU_D%Ԝa+RP`J̫ĥ*14++4TzJ1XK%ؕ]#=EQ7DwL`ƉԸ$m&Sl*Ek-t eqP-=U#ܓin բZtZ-* bZ (4ET P1TF6λzjHUnCnKA(ǂf.˜]"oa8VY4DpKYDH&ϥZ i-R&0FnCX_xz xrL\~ސ?kP(gK9S7]OF&dnOLPnO;*>qتAr۶nΌ[k "Pe)قaZR<2Lw?[?ԍvUu3g U]fpemu˹h MÄj̅ܖWTGokG(){gSI`W'NvuRSYL+i8>`}<8zZ-j hYyM9%g&.`N1֙4eHfi@s0R_<81[n3#)%!S%6E*N$I`'U8u-ZTĂo!8[ Lr̐eLy5hBARm%ҁЛyhH B$b;/@ ܅n:5B*`nz?s/c_{tu'+'mH@,%aˏgѢyWcly.WZrleմF^R4(@5i>+H oGdع 3U U> gYXFE^+g1.g84tpi# d;-_vjaSn$~sKKà;WʎPQGkFDLdoqQe܁ o:v L W;D\ ZnWg PNo7[ fg}KMtag;O=>7Z(K<57&-Cۗ'Lw`?xFje؟ n^ބ~Q;¡rH8A  οrnL).,u'a5|lM o03oہ~6gMwPJj`vx}ìhz8^ pKX_o TO| %>0p`OëAK).Us%&өJ̸WVWz“}j*qҏhJ'>IRHB)A?g\̤:j 3Xk,ºZ3;Dg|t7/E8]91t/Z|_\,m;Gyb ȈEZ=WCo7 O7B^x"7YUFO*)im(Dޕ4;k xJG;q~I+HsgN9mY=kdΌhcn\zپ7\_WzK%?,l?ufc;> !IJ+I7_vpvA"t__~n_Ǘ Zl!˜?^*'X[Pe TLyA5da%X$7caWK ՛ƾ9l3/7vw&z|[a+Jү;HJ4~oJ5M.}y^ varϧLگ ] )P5Jk AaD}ٓiM2MAlopBOk]8] N؅ L1ɲRbBCt7Uȸ`ru mV_ w|~[,O&6P)m8?Ov J%Mid0 kmL% ob]}p75?NJPr\T!9o#D)Otf6VHȊUWUU8~B\#9+μML>`(,}U on۽#ó? 9Oa#JmUpQjkholqHϭxTʛ1[?ETT9kvuoi3*Isl3,E]5nlngby0pu8\..b8Y|XyZ1ݨ+=uQZm'j~X4]=eJ̏Ve@r~ܡk 7q$g\CWKl,+P\;:Q7o#Pn/$~ e3fwU[x{=jN`v0% D,Ki%V,h;Z7n9?>?KtJܥs 37=V#XaMU3@q{9¤v{׻rq ?`DRT7z%&.y2WnQ)a#jg6&OBmOva7q$W6£ȼ~?E2LNi6N#'IZkvPb8yϦiVm[=oC͹n& ;tˉxrcy]Dݼf? , w:<p>5&~yDV=&ZWR[ ?/_X7̨?v(^U5Tg?Ii"0?lR: Tӣxr&U?AICT{0"HƤc||/uQ<Bk4eq/'Z9r0}M4r+!r+C`ք5X9C#1cOp nօ\"O(N-jW(/de6iÉB AnAFz![j7^BRq֧Ĥ} ;]HETw=$-H&R 3$#e8qXkjŔi=Vwҿ0J>zUNrFtټT0GG i3!|ʍ˴ȴ1>3DP.T," bFԊ %Qߢx`XhI- רI*qYo HD1oUA\-vj$JI+%Z)1ס W\`Wn܁j\U BIJA`"& AZ<0vIn`a4Bh-/#Mv,<Ȍ,ܣZ$6 D@רmKQcSBBʝgȀ%$6:uz)aI@#*R N(p%G2 D0fM*S` Eh>iv+7Naa[k3ǝtOJc )fS@[ZouC/+hF,u)#Q3٦6S{pӤHF?Vc{+A 15U=C$B&)T:kc0€4?4gAQMsT=\w;@fO Lz2mZGʥr5lczA*fDkO.Z r{9''Ÿ5u0;]ȓ8v-m}X4BևC~LcZ#O'̡eqGA;ۊ r O(=lGɘ>|UNxyiIr{bΫ[4; ™PE_gק)v#cS\x5?DB5~?/yuY*v^$=/5E@CeCBi̍-|Ûh~/+}nGv̚6Ux^e*^ͤ\ :ȒFsL*k:EJnc9 Ax ]֝b|8|>׾E,~ ?@B"|*`oXALu?S..as5&/^q#K|Xi|dSM`Z/t5>Njc"f33&d*{au.wmsEf,aX,_t?}+ y\1-xʗ>z:JdbD %)':!2'RP`> T1SpY! AR{ V`,0bb3oȷfJv&>`N1pR%Sf@qLLzS re )aL;Cq$FZ.RD%&"60;1I)@(3'vzAʔeRX"SU-窵%HmT +9g\T0&a$Ze"A&f4j!Hc= c!#(u&8R>j)k1NHAe,͐GF"v+R|0}G ˺W0Å.S2SKhA N` 7"A',ue a0~;>t$?;`HQՒ2gf@j ZuWX"$ w428G~f (P vo 9@t˸,*tE6[NY\D qq3`/W.0FauKvaL[wmqUWs|޾z={3~%7$/g N-C^~ބݘ! D1#u~I'Ɠ)mIAK.{zf<3߇R D);7:j 'EgqLZΌ䈩Ǜ`z6,5 ORzAK!̖Lc MHOJ=MSuBEBp{Jqkݞ)zw%.e-,lk OK^`M Jͭ[vq94NQۣh,\Gg4E֭h4G?7(ů, :B;{s?(-רjCJ8[V+8BIv=W $gfYTRJf]1^UbУJsO^O^G_+++qW@f+1#sa.S0BQbNiR.CJsɓW&t~]feGhME&jUg/)ÇN:׺ ZW!^*k]ʉZCyP8QZRj6a90> Q:u#1 щZ뾨Vֺ/YPKU *MSדd*`*d*=JORS De$LX(Yδ'p$0+W4z2>˙ga*bmwr[|:C0͡o A?b8/ :wNz (ia3 y҄9%#Y*!c3 -8^)tqYgYoځ91P~,=oqjůӪx]i 1MTظum]C,9YH^JW8E Mĩ@Qɛ:bw2&洎i)>ⴎ@08x->SR íQp\{F ;aњۅP*R㧬=cH5!q౨RՃʴjXTL*S ks }B ^R:x\ozI(WBOSis|RtbEBRbׁq`2-5GNgDq#KVK*uL] LJcև{,TbqV[d"Z0l7Htה+ZYLfJkVűG``^Xxꋞ7!0`4|\5m7fggqt|3< wlj|Y`[~P՚iWؚX'FW_s!L_n,"(Ȯ^!0.ހ MQPmISF$s3 %0 2&Vcb:#)% 4#YTU12r̖p8)ӱ l1RBc (Fb:0~yu?>jXs%\ pFÏjp*fXolp\8|HbZPSߍ ց>F_4mX` 57cM<.nnlpj5:Rfv zy$Нm-*6' ]4l&[{r~$nblvHu]"# G_ Ě3;5<ܜQ"~ͯ~JIݼKTw8aF L%(1ѭfR3s)Pߗ*IJ|4a2,@i@eaїQ$lr v WCfYK'Tc0B*|_VǏy\> ̀߻< _= m])`L~nznn'iPC ]3J=Bཛྷ뀅7~q6\ ^OC0둱nh(fH58>ٙLu_O-^H3M|&Ԟ[%}%@yI#{#fx8G175X$~ lkU)j'W#)9+ Jn/:NH9 ktGpXSuHul lԿ)yTm#E9@K̝rJZ#r~ H V6}vפPe)t>Akѽ(ƹy̼L;c|=O%tÚ I7S$E5 _hKbMDq? F U=FuAmjsZ 8nJL?WHw3c̢`ެ%:%3tH#ϬR;c[3s5抲gboK͵փ V#H8_?ShhWZ\UAAN)tjN{A$IT*Z+5 پc߆W8jq9Rc[`-.3TdVJFϭEbM_nb%6uT܍rWF#?{WǑ 7l2z؞D|'"}Y]G71X_dDddՈ.\n^%WͪjNJu:T%X0{z I.:I&s$ L+Ťw)rUg ~5X/ B ;E%pskQ5h>;lVA_%zBlwskٯ C#'-4U#*5nT> 7mT^´vFe|6ʹ2}(25FlnJGnDtF GyPq9T08q`7:qCw;a!"u&mG7VU:JC#?&V'WR%_I.9\jE1dBx9j#0kZBbB"t\¤2hC VؚWˮ]CcaY[ *۪BC1JoH34N9U5^qlx RT v c[xkC;Bےҧ"*!KIAr =Ս Sn[:s!=Q>Wx ՗}5u6H*i8U "hE29$6 FF-H]EU2U#Chc֘Fgl d^vn%ژ%/Мc)F1 GG1*'@e~BtabBMugU9)RevCIB !b J$f)KΜs'Ly?{r"*3r37BOE)dgnEX xO^pVt8cCnFa/կeEכ+]Ί/M\O2vy \[ߴ^=y~>=Owo8#D7F9oȾ̯oE7A焱o]ם];n₞/ ܴ"@m}ro` tao~= Jka##2yCXΆjs^i ʣφ,{x {TT J(Ïpmx'zQeЌ2Y!mޅG}ܾ~ff+t+kfgN(=z,ˎpvY2-&7]R@}%Ō3r&s|uM`NN0öUwC628**Vs=_-V__Q}PLVݖo|HZ1=ր<:Ź^JS ۅ -:GljjZlH  ,m<N񔼆f՝c:&*,= =6 4ᣙYCJ3PpiEc%Q8VXЇu %JRYU\kTqUtU%p 1rY":1%h['3"rÜ;E X>ޱAd3^`  tR$ GV6.k;!;ń!OxG\U{nvВ ] Z;$JLz2'"쓀zɍbX1rW]DT`p*pV/3kyD[V2t>kOo>sZ }t/]:.ٍ7?e ޫ۷o{,qA;td:@K+ž:fVo pȴ}uޓ.ý1A eؚaUKl~zryOi{ =4pWV,WLZګ4"ཙ3 ;6#>Ӈ}tԮQFr7(uI±Π堢(2J-XL2X K!sk'6ٟZ&pd<6h @/T$cʅB%w;bgՆC Xt!\lx;-*ˆr+^f`e&E}Yd)8MisȀtDFyBãeLYɂ(ՃZjוzYMi_s" jֺ–S{[@7{#~9vskC-^}<]{sӣ;@#5U|gUC/Áǩ<5 t"3-d jSP]dP-eB簠@qD5^b&Qs2B9=dju)o,Y3ePX+™(B.Y[(j4dSmѝl,8%vE*3QP9;ůs[}.,4'ZIuYD'äw4*Vm2h_ѕ*A'vb՚{[XKQhB䎱6b\AtC:i#1APTe0B#5$6HχIZkܘ <(\pN3&_k'e+>\zT9cgnҷigg877}Pӯ|(E]AHO?\/YDHZIC@2ɱ6-@2`CV/tZMsv!b,-'kXeO\b_V ^vː$hxϻdo@Ki&c&X<|Cd~\0ƪR0V62'ָdu#Nr&O&8Z|VyZt =zzlu_^OךUXoF׎a⸃[]ٳCʎD|~k@Zc׀^/U7/_*kZS>_ vō؎Cu=g-UwwSÇej+\kN 8}]+Neo][z=S-08a#OOl^@:!ja!{`ӓ˘QWv"Rn;S*:ngb8؍}raykgX0CCftr r98@cvL|Nұe GG|''cu㲤${T&H&tO {詒GNiϣ~Ĥ 1xv^k9/Whdo9c8fXv`!pcyG84F Ɏq2Ri/'OπЊeclɨ0ZebFI9jLdUItA845T@#W%u"6K/YHR1SuHrD/4yFK|u[&X,Of<F+wYOKފŵke#k{%(>50ga~kaR\%Yrmp_c\%C@rA6\#ƽZ 9a7ې.*7L+L8eD?pUW;V[C4e\ PWv ͗~_ N67K*H|VVé8ZSH=wmmYzYM" IfaukTx &%Ȧ]jv3_DU眯ΥY%%L} l4d%}PR M3>n"?_XR&%pπS]*Q@5Fc NH 6z>O )y| ;ؿ";S=юRW O/ Y,IN"%v):%vw. ʴ CWaj8V!y;KB0j^r$1 /#JCeћrQ3_47mfڬF 8pe$5@6GPkt}7OF7F)uяPҡ CCџO?P'u^%RZ,F06T1x8'8] ÅC#xe&C=,6~2 6M7_af {mX f{lXIUtǸЌ؂1* ,.$!ܕP6+WR(&%H<܍bS$\(z@mFS#)\K+8_mΖd)j!~ 4@F'hK0/QWRii9QKd):Y:ֺ6-TtU=U2ЦTMQ 2EDMNpNx8u`[ nB%yARn9Q#'6ѩߙuMRs3sJQRpeie):$% i+A(AX`Ri]@*U ɃDdDyu_rCF8<vQR\bSsy-HHӥrRl8f\RoY C!{m~%7VUD!=xŀ)I_c KH+͠ҕ2S9՚[WeU"ܬV B5bWWxʓ,y%3!Uy+|at6;Aţ9T`>&Gmmj2 yج_]}֣5FcQ]3[Olvnb~.=Y涓XŢ,xjQ;YsVBU%CLFM7qCXL8}j3{s՟NDv0eRz0ع0Y/MdsogV'oWvrRUy[bcyz y+|/B)L)CD؉Mn0o?~}r9׷ xwc8ˣWE1V|5|;~j! ("t$y!%NosUv`G4D:nHCϘ"j;%]=wm%y@Ӂ)vH&jqi4 ys! もMsp]! ƐFCZj="C`̱ [CQM㆜=Y0scx5݄=F(=<=QI@0 #'aLM98*$&ѐ9gvnr]PMأq@fk="{ύNOs y4ki<yLAgJ\BQ(y:܆U 4m0"Vˇ( %㭑ۦ{JJ8 % yɯF1'R8yy? &޲ӐzRq< CA0ȫv e\[< Ճ ($0.Wۆ{jʨVYN/(RDa# LS@՟B7՟wQCKq&F ""ч֣ ! qNu"t܀ y6IyIv2D".?\ov&>? q*?ekdRΦZgf|ݱ~j&}rϣZݼyg nm?<5i61nvW_MGog|s÷Sw39GZ]8kw+6wc}J\uao}u[qHM cw8jwXj;IfTh%/<gɴ#@}̤<|(zF%~"DFRͷ7Ys9LRu~+婇=)򠺮7hCE՟@J5{SN Aa]BI $${-:Տ.{Nqj\mϫpHX%miA>8 `eMψSfcV1pkFԁ=ಿg]g)iDw8]<5d>]F#||<֎mOI0<!p |v 5v s,&TcVU5CWEd* f$zylq[=Qeq]66ڝ FX.n a5jG>v̏ωO:S%}1Gmc(M_UT`$FѦpgv*J^MLz, /)6 )#|&SC[RF6\mJeo-[TB>sfaO_wA trh݆+Xl?-hw!9DOƔtI4Z޸;~ׂ8/u&_L6*Wq;#L(8aՌ(o(ZM( Jq-I"o)Ema8݁{䢞촓۫{bs7⢶T \ެx..H#wN5j:lЀDӠ*d .j|J*d+ݪDT׳`by-wѿ.šrofy=][z%;/nnnzV׏/$a]E6Ö mG#_ɧePF?N/Y]R5fMD&ʳ k%v{,iĹSR V$&qk4uVk(L:H)0G o~<],FkGRK&wbPxsQ/d ï;J7)Y*=զZ()e,&XZMh)<DZٮ L` G{.rܙ ˋwo{Ž:HyFV)*}=OVvYbwSCh3uDRlvڮ?m<C4 SB^|yݔҗAI1V<kG+h2|ÔLĔ#dOg\|"jۀ>IC;{H.<%BQ ?r_+B,碰9k\YR\b*dBy i<`E3%R99Q \[V/D}|}ULBPOTKNS-{5ZLV%T]2IGfPo,b]} V9'=G%s0Ԏtq6H8C>w+6w+fYD&!XYsKVV+%()S yLn\,Bረg4N,NWΣܽV*`Lܽc S>EzV忳H] qIET}(sPC, bA`À"&QxI1{t!ozkiRGalJW-?I|bB>sfaCEAru6Zy.#ѻ :@g,Lz8[ܟ"UAҦUmWm*B֏sEih&uL Kk,Gd' ǔ!;E ,YXfnnBIWiQߟ iy p%(8L+*LPyCK* T@%XB-{&} 9KH\ Z.])_(:s9"ftB,؀ݯXIZ@'19 &bLZc%ƃ'=]ihCAM%1DRue^ii( *Um޼4G=gob9t5yX.~y8`1Yma4]6v_?]ktUjInоNhI$iRE4b^~\1F\c l_1-of4!_~EqȋIz*5G*Ŗ#|e;:eM!,mFSVf#P,w^L^tx2lAY8R\e7H>%~ۿ9Y1 fѤK<籙e.,'gMkΆ)@r$S%:/N[UapҥlҩbFJ'K!I'ΫS_&9NMRtD+¡Icn≨'5Q'e<D a.Eq[F׭*FDϖftEi 70[)2[ᴶqɐKN_4[rvٻ綍#pM&tnQ$/ms;؜P$KRN!<$5oq6X?/KTXUhE;Vj&we7,GJGyc0saՑ9hdim*y=)+GMcA{ڏd<g[ǖ[ӌH"Uܧ5~DF3o} JPRZM:1 [E,NI1R)JT++z7MU`aØTY(!S03amJKFQK\*%D#G]-Oܿ%)w S #y)58Q &0W]ՠeu3}|]=r}n"7ǺvVTcS9\ -h1]/~:H?_ڇoyx0@o8{yar"QtR^j?/Υ)tiq.MKhn }M5&#-aͪv>W(#%)#g[1?en)тKZe*btkJAJP Mm3c+S -h366B uC g@>r"o3\`vtָ*.>OeAܲW/ܧq ٕg7{{5fi}vHt vcRKw[RU[O}ü3vnpFuFuǣ(/·c^ő/h4GM=(8FAKHSAk?MdQ?7ټvՙ@ T[,3Pܶe-&D˸ik76λÝz{ȜV2ҚpF 4mU5&jVcR~=ᴞ#djS[ %fE+\*v`/vI=lBxJ «1LO+87Ŕnܰ{iid[ vRC[MB>(jM>(8rU=` Zp[z)Y(ҺOaFF&0ߦ|ϧ;&{ۙ ~pBٛ#kg<0~_FޡS&}ʱimLA^?7pu޽e4u?HKf'ao}~d?#5vݟooOvK7 ߾z'}Xn.u ?z__fzu5z ԘGV .w]ϙF.ﳏpנ&%Rx"%eRɬL̰TiƭVc} 8Q^Pνp J ial]vd(3_N`rE]0-/t6;O=D{+3sL,oyH)hjyEpO/ͰݿWotg}ل_ Yva/r# :GgLzdff?:__ `qQQysN>YMdp8%Ri"eʭC9-W$1RQ$ :]bٖHP`*ZV#s#8-vv-(>- rw,8V}7޳h0.l ;X#dSFQ4.Ƅؚ&A3QoRraPa-F~tu4=M@1(s)^[QX`1F!̕sES O~ksWJa3E7Jj Z)eZyɱA[LEĠ)x9Kx Ua $@r;B14'G4V}:j0h5l\tWN:_vW HV <-4~Oux}Zj:@P?<܍, |6oCieL#{Q2GOt3,c]ozc'J_˥'0?y\ꑐ'.Q2UȏxS)Ki )'9/&`ӹsU jKM҈Qs$VɱV Io_Z KAiJ䘒(oΒDe-I EtEhI(Z;) idR%Z!BhCOW+.N`%c`U!k-vay\\*竂J u;<SLk{ -Lj,eV{w|SR擅r7/({KnY4&X4WpKw2)t , 9$[ A`:aBcބQi0#o f.1'u 6>MM%"t=)eW"_a#S *xEEZir+#b*e)xiK n(O-yJejTO#Lw'S)TL[ft?ϖ0[hꢈcmL`%3i2絜2iˤ)/ cPTiRGM{FW?j*uլJla戈=G?jӞuJOZD LDBtJ(}d1'TZ)tĨB~NjQsڐ[AwT-?/xʛR&lnyE4J4kݔhEL,eD'UG4"ihBj:$䉋hLdMb&[,eD'ULibZEZV!!O\Ddj_߭uy;"H Jo= xdsi,#$䉋hLm]\nvDAщ}Gvuklw&vCBTq{!M!\R:A)J"9Mhi߭ y""SL3(b51}P8t1&Hդ3N>f?Ԑō,.+zW$;pܟhi _tҔ8^,w!tw9iarq%FEDBbS+Km$ڸ!k|9Gb.,*1\-rdraebeLTڣPiTܼ8V^S3oP'tXKr-;̺83pf&,d FYٯ*+下cBuF/{y,?hv,0T)T=HR8 }H3NHm1RnE ҍHuT/u;)xWӻ.;f>l Bߪ:]/+̺a~se![7Cp폦 stBs&ۀ&CCgH V")hܓJX)iq4#ha/2P&:1Q .&>C ܠDP1\ ճc*ߍL2ߦpQ*.IawoPpٛ#1˫E~$4iMzw0C'LcNgX½ ~g_o"{{h0K3|}}?y{S?#9|^x[I6BN~/~:`pYͥ?t9H6lY//i3? tv=Awe>i jBBt/4}.LD O4L*"T*͸Jb,' ʹ,H\eydgal]}$+b|9bdzi:e\NTxf QsL,oy(TZ!!5xzi݌}?(w&\b]{'lkVOrCYf%nbb` ]igd4l(3V0Ӛ ttVb|Sٲ +!Agi UwsXVe~S2-vTx$'=)%1O|/N 0lC`.e=ZDԋUB}Dri;}yGMM 3^KDj,1 qHtD2 ̧FSq(!Qʑ1FBE ѐr(FVC&$ 0IS #~2' &Ƶ `e-<q"DpZI@!W`MS%Fx7I>XE F86VDQǜic,b7Rt$y@ f-S: S )S&O#*4@3Al k61E,L C;Fuloulx" |GėEhݸ87t5n&|,V^r6'/8/1eZ݌W`5 .0weط 9~͹C)C,é(s* 8A"49Bq(a1Rkw:q)t^>1'6B"A( V 6&c{YĘ`Kz>C4dCk(oW %lU\rH0]O6Zm ξj3Js~08à }I,"Z +|1;+ɕeAd,G̲5S*^y Q*53*^u<tEO^rՔ/f_XZ|Uec.2UfH<.Y@gݳ\L'Ħ=|WMdNBx W0 6kU몹DE`r9҃v?6G#f+$TY>D>;lXy@d#P{lUWFrhzʯbpTrn_ )K5cF;$jy>J@xS/9,mz+_3`SpB1yZi v~w :`9W␷9:lY˨zn#Lݥa489QmYN_-;9@sZ`f984< FqBywj0N*=sT z_?݇hgLj8Ѯ,8DΰYG!<8+}.캂¥v$Nk˅JR)i[z]4k'q8f KRhԺT&WRhG lxVyD>dj¼OQvb!҇]KO lWDyQ6 K^|j2JG.Y]G,IP[a mFktӴqNכZDhZFp{Dw^ӱ 4i6v[C\)ٙ^?ȥ56.t}GC{xiv)t= xZ |zv|5ˏ}2]ݎ[ lxTqߍ̚-4hdgqѤtʢ]i?\DdJ֔) [v+ɆdFhiVY*Yżi^κ\w>%P#B$QQB1EK̈́[<:P@rߢ {ΡI;)`IذXK$9 l()5 a%B2i/y =%K'XgY+]݁X52:jf| =|ʺ=f)5v-f>nj`/}bU}vIcݹ\n ȆMӼSz.7~w{#SRk뇝&ڻwiSc" Erׁ ˷ ٙ}fCvdXSbkbx0Dg>M׹5{P5yYRHOhx1yv <˰+|/%CBϔb&Њߚ ߸t*\JD82X ؄&D^ S,Ej3-&"v`A9yխ1{ROQh#4Y&gbӇ"h5IO}6i<]ΝF~ <- y4XGmQ$, QGI+mM"0t{:qYsq=)tΗ#8:@& e {E+uC|ׁH#23ׄ; ёM(HaO-+s8K46VPyXS9; Fif;G)v@ۑ{ߍ~ɩV7aMX1&~ᲈ CԭSjG~Pf78nGZ?K.%岓eǷTBxRiǿlݩ3β-_Np̮ĸd=pAs8 Jr4#KqsY"z(}׌7s9FJ@> ,?_fe^% 3X&Oh>|§Ud}xkr~ғdlt%e{ NalNÄi*ctb#bEAv{jy"1fjz5ͩJQ^2z9߉=$iYϬR~^mI&W 4p}k)?:gM=!:T9UFv"BYLo/zUjPN@G5z#Q kA;gК + !5!&cD ěC18RE>&z0 vX=ن9]:W M}$#Q4VoP#|+?uvRv7 h@Ơ IkBOm+C~6GTAwo {bw HT[aWK`u?OQa5Qa`x^*—u5g}׺t6`6[ݑ|_\O̚Qjc24-zN)~y-'"#;"'7%ql%}_~̴F(W5itTOa_#y=wKI,; )(Aywmۺ) H9-S?s^w^e;lQ1F<2z])j\znKJ)KD)oX 2 :^yg4B[Ĵ8^Tpg |nM?ɧJ 6^F"kȍ-c('I11pHP=-9՗+Kїq±Z_Fxd}HHAjf˩MPR2ђS!J 2Jw-ѲOʸӽmK{ݼ:"#⻶ݦ>SzHAG^Lk{R\D؉JIj0kP8M+[,2ݯY= O\''r Z3'rCGK1 |&%.DK7:c5VcƅZVv5p ! <%nTӴN8[rPN'1?!SBEUq R0{ JWRJZ[5c^˵ ꥫ}}Vs{$aE#ڜ㍮ #@CrmBg`FfE+|LBCRB~9x#PddN[qD@єl>Di@S3R?dnExOm]WخʋTyzYNO`K"}HGnRQ{9Z q2B O],Z)>R)Pg[ǏTC$ZymRh: F[I)դ1Lso+p\ 1cZbIk.CE74rYubQ4*OzN}ٮ`SO򲮴IaڃcW$\ryӽԖD0 .{rǏJXPD<%?=j4\7ACJm>{,YS=Z+\_VNk&@)UTBWlVc'ES򴧿rA.d=f \I.|AKM=E7ʧq GḎtK)tySxb, @')~e=)ا3u,3Z ǂ(4a{: ՚xAW{s1viFX}8 _x>(-j]OTI#,A&M;d2[cb>VaFsRjܜ09 ?gduadF`Ev4r]϶cDB[C7-vdNeH .%*DEІF!YBy Qp*Zjq?zᣰػ޶eW} Ni(✢@O4IχXK[,)-,)YԛE\eDsgT'cɰ8 y a ڙդP/NC"m[ؗVS˾P A+-=`E0ql}џNQūO1=Ca۬ģvor7`AIEx UU:k-}+ꩬA$J}ʨ:sݛRˎ@iwYTn$k{Wy$8RGKJ5vɇD3{kX`Mqu +gCϡnYYvxAm&IJ!}w=fӲu!q kT˳QTvgth q$BAJZ9!r)z&Z|ʴV A9Q>E|NC  #~z6 N7gBPjoĈs.nMwVI"pWnDsc>?S5Q-ZK86'JI„ϔ^4kJҖ4x96"^2S^`.iWpI=^8 ղBzn\e{0VPb&}5-{Z ANz(& "딛S%$%=RY} 4B}Y!~x4J؋ߍGZ5#SO*^(B\ar8TFkE8xj(&$r< w78Btw7qwakWs8#9䬾tlXV58" #N~?|[OvF㠺<=! zNz{ř޷Ec)`itX"bSZb&(Yf2&V!m6 т'r_`fӲm6 [g)zH(-:z誶lh Rө-+n=Vg"}S R\s)8Ϝ`‹X_@r3:/]䣛]OM^;,yK ƭ5-TXP~J}A}ds7C?{ Jz\way8U|@fGm|(jpq-K, ?D<S3LCpft4&gOf&nxC5 mgJkImJj ǪOdrٛgn[6S) ?'t3Կ۠^e;Aލ-^HɊ{t#Il, cX U"%TPJn"R.pDFn_9@޾\ 3u$\rRwJøS)q a0S"DRa͟*qH4k,e8~TȐ0DQ*ph& Y-0di80Z#~K00p~Dq.BC 5({I,l5W߇ .Yp*e H@ .}k̀c 4` NJBUtIwmN ҵ8"Prܼ8^*vvׯf{.xmo3LIj-/sF!؞dA$3OR 9j\+P=oY{sɄsڒw3EfPo5THbѶ7+puV,n~ nFzª~zw`"fu JpB 􇙽7v2Son7))?`I~Ӡ(evS`Dri8(V]%N!Rd z01YYsڒ.{m4C.=zA,hOk,佅됊9ǧ29i={q;7o\N0hm\'Qu}-ës@LJ^*2vm_G|/kp 2`:=xX4'u4`~#޽ ǟ|̬?om]Shz%tuv+o{7蛠d N뢘0O׾IyTfgΗ8i\o^TrM%pzRWed[>u~̗X\}fr5: 86Yf$L.^,0qɊibF?,a8\vmԳ}jo%_`OւJїՓ݉@ZyJ GhyT*mx)AmpnN r"O4 ӛ\n 4l31{j8[ ]|NZ9y4x6NʝuSxy'K'DelG8&fE+p{'1"_7\7lrVcyqG<;AhAZ#(Qhf|kW;駏o* ll>?M9#Ra"1qhXjCi$a <;"OK*ɍ8@%: [}uW Vzz<2 S6d426 Bb*Ъ2V>#i,^Y1^o5xQ^-+xqх'4Oa!N!8hs$QCPrlNɈ;XWN dqyC8[٢+[>0̸*x?`}lө9ζr۸㶡^ XRO =HVzD^-narOTcZ${~O%'v]A^+"zy]8QisU0Fm#v`:<}&6 l?@A6*J `4'7]iнՃL{;}B>xHu!R^PYv A)/GNiatd5X0>yY! |i(?E{Fh0?| l*uMlK>"vv qӘͦf܌ge{zY68?D!_(5 A` G/NJ!nKzr( QxYgEb䖎9{/7fDZ-&`CbvLs~ً#G&ORQg!OUlHUo[.{[<'; 926r&_ڧ&Z]aю Xv]2p9Ç@ -vrc>?z__:Gg0n-^i ꄝ:yo.$NuJ,CZp)TV&,N6$ᑲc TGǒ1A$7ժ< ɛX<6$x4!=d8-d8 ݗ|xUxu/>~e`C[z8껏`E-_`6Y^o+3x"!\kl&N=,}jL|¨5 1GFDs B#L߃nSzXk]0|u>Z`5Ȃ޹9oy:g-~ ?؝_ʵK\ʵK\+ē(4&\ 1%L  #`#ׯEX cr޹*X򭎥t|ǀ!V*MaO]k? Iot/im>1!bzǸvykǸvyr,i(5Dqɰo"pNjT $2*)I*u݁jukP P^ JxD|z#(CO^Ϫ* oF!/97ԭX_֟ "D"4!.]ք :VSWvrRdk*8Z9:K'uW2P{n6n/̹(81Ӕ Cf,2汷rRI\`SskrW~=vZ<0@ߜ}\_=.ëOwlMnBR_ER| ؜]c&;##{BC+ǒw _ :B0}iP)ė AErOY'ɣDZFc[&2TWc1&DgEY^b(ܴ4&AQy|Ca[fe\I!{>?K (HYP*Ek#4bhR>(K鰗Du|g*RCLJv1[dF\FS~vUُx=փ A/Of :};'833*x6%:eΖnELt1!wzP\ᐿ?6hϓX|Qi~:C:*)32 cFkjg<ۇf}]vfX9ƟdMFJ!EnxM'T#{~H>L+&L 9S ShBjVHa) 9g9UpA00><G8Jxt?46S[3u )BiKL Jh8/^z,YX% MHuf6m= sBzrla>+ ,zNs!ezDy{ir^F!Xt3[ lXWg%yt&2]?\NO@7g˻s5pewrL$ڳNuV|XF|GXROj+`!Oޙ|[8Z§QozZ>h5Rɵ2/74\- QGR %٭¬ ,\No:+SQWz;oF{Wo:5K4KlJPڹ[WiSM/pTw$Zݻ_~ # ؈R-i[# :\Ǟr'|$:ntg"yyQ*0/=׋UwrT)yd*@$V1BUjEr<{y9OܝP+ҽv"x<Y_绘Yۼ b&s^o?iھ˔{5ZG_9?;KVP9QrwOzjx 3P?R~s'LPj zdCyкL0{Ȯ9©W8Pb#U"RQ2NpBzh4"*N%9h:F>,Y6uaY"bH)u0=/Sʠ@G}/SP(P>wFKDDRTXόT1}4{G(;GZh-{oGf|fDAKsѢ p새D8sV*Hcq4ɘ]( O`=qԹ0@2ji@;-w ΀쌷XaIkK!@+0Vc"kYIp:sSO#  a72QJ3D%K{ SvB*3mXѴq7[r7ۦ^DzdقpE$U?Vm+fem~9<ϊX +͈Ju1MY,Êpp #WvwF"H|<<bS6Bu @jr{s0? 0+Q*%#I#k5AB>~p5+(iM# kMYdeD9`*P&RFF81X-/a1240bHsPC4NKaT9ӲBɌM)R4, s3°cT)U)GIq`heD#ܥ3A("`{jP )@3u`RR #v1VGw{ӈN%;;+c%!򭥿f7=a6o~|| B2D$+Vsg9xfS¬j 5{*ЏO cڌ_8x% 4t|BeaO,R#L sB2%1ՄB ;R`C\XI=wCFMR]s~FƵ| ^ R WxyD۷D,Y"^\E-O{"di1lQvԨ5b^^qumICG/wrjCGQs"xtNb$lt7(N8#eg6Q&X&fTOr|v$eIʬ˟ md0h׷U/`cgw?jxcs/￷H#\+$4}b57/ݴuqxܧGi$#qH׾ 9CW.d4不D5BDZvFYn_7 Er"GtPyNMN>٩umͻdփC!HB[ϻ}vJ{mV_`λ$9l39wy)#HH3ـpGMoQ 8f-炍cv)sm]*s=J=W䘲Y[1nʬMZ4rqN7'H VbWAv^*κ)nK~cT4bU敬Du1*9RW k5_,1}NF/^I5&h%լsK2OFsI{-+յ$uqDd@ KƓh@DWnywqJh?&\goWaMV#4>8 ړv@{|Y4&7]>,;vz6M~{.B]̯ʬbcy6(N@QU~"זQ} -0$ШAwpF%ǬmK4<=N,Mn2~I։Eoaן᫫DB@~L@ۻr&s8"j>BD %.HQ]W~#~DK晀sBeXmЩ#)*XK*8,&1Z3g JiR[NTSt]ڢa3& 3rxqNk˰Y+ES4U(xϔ'S8T˜#uXڰRp˓[3:9Њe~o(`(D1 ^t䃳 A 0I \̨+R I|qT*|B1gR n(T+PD9υ†x mO}ڳ2DyC\ؓ6]LjWⵚ:gmЌo+Vʾ;hIkd[#c&  Ej) BhBQC:МC*6Um׆1>f~-8asSZ4A8 zr* <=#5bgѢE8"[GN5ܷf[Qͅh x#@6<*0sc彤 k "R [N5%R# X= )BEI`!$ũ^")/6_/k3S|f܇շv1/l3L O?}kq oț_.X<.h~ R!ZG<@Fgݫ8W~㻇+W?g}S͵c87+~\vM K]_)H'3wWos{b.`7C|FozӴ';JHR\'7ף" àw'!@8Akvd5J BoʨA\J^Xq+7Fǿ^Ƞ Boڣ;:"DE<5)Yրۭ4KB=8'@`Z lޕ#"e0Mg4A;X61~GYuSudISyH*?.U*"8$dY BtV0Tc*Mf\/H|=sfm{κ] {ZRL7kMP tu# N GcE/-+;؞(oCحf̓Lk>РKJ4}RG Nr%uVn' y7vU.zQRㇳ`/BWr~nTz{VW?YfiCʪUy}3v'd=W9ox)o畭>F| ^K{&>#c缡 }}^LjyC-*{`.ݻi?oJLdU>9tiK{Z7ydvqzcڲmNЬVŒ&}rW,%rbyU|vH:6cju-=0d+—!]֎t J93:l"{]3-/\Q4r Z]$l~P,.,a>#dX7>W 5+qu^77|OLMZ44{p*;et"99Sxס7>|=/t5 K: KU^Xbڤrx˫O]"(bSG2cuzYNc?DȚJ_ο9åh=c*("폚Լ?nZeSf WFndˏm/ yY2mydE&T켆 #5^|AѾ'ea/FDI4St9_5%K!(n8x\R0|l 9: 0(e 4 2C1ʠ,LBh *=aInCzN*HLpau+ě5ՄS)2\ļZMYPJ"˲TY3(B :|@p^ڲde -_M—KޒN3P˰VÖaYjo]0dZZh o-ro(܏vkCŁ7ۛIׁL>T'{%TeY9Y~ A?ϖ'3y5Й¬%ԬKd!gY,it䢏,IgFyfk, e*,zqT [:BK G" G0`:Y ,t[fӣF\J"zs̑| 1"O|IeZ8RK0h4"8혶@p'{a`~B`kSG 09II>n™= T^:kO.0"=%Z+adKU":NdtS@e@CY!n|i61Y b.W[@2aY'~qzց 6H'q!J&R%-,B+5/l}QFElԔ&ތqw|5O1gg_!]Zu#EM6['bn%OEs ddˮ֖7I>-,v~%Wǒz|RTrO.Љ] oIznJ~rl>zK4] TP"LOEuys\Oa^0#% cμxeh9[ЈgRсjI^ؚ_>Ek8zqv5jÛ_^.d93N1ǤA~Ɍ(%gSTt,s^hfe#\ k"P,Ѿ,2BDj轢+r-!zs]뉪,vv NJHwM0u$.*#b(D(e<NE&8kipZ.( :\C{w!L4: cPZ) (Sc`V)J Ŗ1n]]I I mPq!L-e9p\n5*8}jL GJ:7JLe(%%(qc*f FFdC$y*+&r,^RLJ#T FjFR E:!()1#y~Q}⮼L(z*|JCש&ORl$&ORlשqb-A(c`48D-~TtDg,ҕtk0gשuSZfC!G}Xg8c3b!^$зImOn{vۓn[h04{"!F-R+MpM1\pfdkr~:-]$]_)t)GDr(k)UгGGF2I 6B"!PǹƲdX ӕ1<j`[5@vTCEp#UtQ;w1S<1랶#đM9J,nrJ#'9sg;tAkG^m4 .x|dWk;CE4%;YX xK^ ^1 )w-ƃJi-rLR$g"Rƀo:!T5[Eb$p׺A ]1l!եDs {EPrؕ2}y(URf bW6VPzL4FO֤bZ0T-Om2  6 9(6<>E( AUEӊ5ZIjyϽf 7zD DtIZCt f6o A֋҈u@g]lI:9P nK p3ejJzktE8`QB)N*֚Zp8+Id^il&פGHFլGe5yTp&R(T骫(cj 2[B%L4) ,lE ]R:Z!JvcNbIfl_[%hjdjB@ JxmO'AnmӘN~:>]?ӀʫŞ]sPjEL'o֊J#Z, HA߅T^]~ďw >asTzO@QSԹsIO|v@@憐]dvŽ'Xj;Sׁ8S$)QY)aǑW߷ȊeQ߾ڳ[K LƐ`ϖN^h1h@eRB\Ⱥ5 X=5O @g =_IF "j4 rY`2 F&[ԃn>atPeآif9+M yfZѠۄ)ԭ`< 2 Ȧ S}Q]Ug̯/pﳭ|׬7afMN= <,(r@NruN*) hcBѣ+D(4RŚ_B {׋mHN%Մj͹!>UYCpV1^z|!q?UdoԤߤSVNW:ð-+ʠ`KK9{tqWХ3S*Hqv8ORJ@e:5@Y^.Z.3$X]? +9_/EқVdY@7 p`XwkpyY\!\j-c5 yu gWe:Wy8_|JBwvU+ΊEBlz,G9jܩ~ H:ݣ|o!WDN|?O?^.WvHIG3Fi!J c(wMJeQn*{oZQ{䶔# -yx[)Gm7]юajm#3qʻGgc4Q4Kۿ1:[O1֮陯1kv/kVOiT;׊@CM[Y<\SB#c^=[_&`!̡>݌qBgf6 AY,~P7;0XxFԚH^[~AvˈA3#cԼĠA}>?(B:﯇ʷ'J*M J}OU6h gA. U0FzO iq)`6zyӘ{0qYM=/?Y]se珷6($IQ[aT~ыݼ0q[Z|?~x,֕)Qٻ8r$W?6%ˀݞ}6zfew`˒ZU3:R*WYzie:I-6䧫hs۳Oi5,-S3~Rn+YԒ$_EUuެؑ;D4=g-+hOҊƓ4IѲ0lVɛ-jMZkV}82kj@ Q @i 8oĩ"TX,oKLrHv2q}Y@BPku[C:N"FBAiJ48!\\P%f!ݰ^^0u547ju׽i\W?M"fDuv-ֲӆQxv_o2ӾCz]] %#7 ~MJj H:\_9r6F=n̵.-`B_Ц+ ci]\cS>WC^^+=,|9q9_ڦś g 7O+W1+"\u ?-ܐC?& s7/"Zi+ыdy9h n=Cl}=s{ d*'{ xP$‘=9^Y{dc  JОq8٠! isBF `R/L^c]DZr-ըgB)`qii_^HK&"a֝k\z4:ZwRe8,Or"#o# nn5CݴjDG!$9V,xN @iNd(OR@\Jgpn#D~2iF,RGe7gs`}FK\fǃrQV3ոzhpwK2:yCJ $ɠMn%;'(%ӹ DX=?J+bisVüe7>'P| L(%xby$3 XuTPEcP<&_ؙ'oy\.Ŵ>^ZdTdX^EJsÕ[iE0Z%W.$ Zq jM-om K9j@j:->D%ML^f"]}E`ky<՚)#Z)SJ1gSB5M1ILA8wAHr(mn`^\hx-JCi!RM@bT)œ0G-sud4zC$XX@|r%8 dIjhC)J_~Khd =WQ8KxJ''%J^K\ s/.'ˡ.%/A[5vrS &fA0 pk>hSt).>p#qj҈FLZV+(i*ͭT GF ;7d RB~fVb%{:!$5t7#6LrT e>ߍqco3'gཌྷ7x[I^уKr7]n-Q@b56/Nfd/Zob0zE`KB z,(3{?YXGm֞(h epn!+pNbRN#7o_$-ǃ8X^ynWdM%]%JpI: ,-=()?yx7@ aN+aICS#A >X)' O,%J9#X6$XDm S\T(ڳdRNM2ը9b\_;.d!k @CY,+SI)^| d ~ `Oetm˒LlyZfLʗxq<,L~[þ쫐{`;mVQ貥t:dy 4x Ǩ#x$`(gY&8(m&ɘfC!v݀TӷQ{w:^)uxFy\:bmF-i5zL~+$[ yQ*6Q|HiAG(I("AJ0IE+ q&a"!F:+ #Ħq O뛻iyRE'-IÙ,hCt5V)aA+7%SO}m_jqڍZC10ʅ⌡6LQP ]z8/~}1ߗ8G7p\7Xpu[+eP !=Q%DQ[9}nUuu#Rr,3FA"+Ez,8QH%=w6DCՋ>LD%k\21-EOO*ݦ`=8P {i@ąJǝ)ܜ6EXDweQ3ZMJmt D+t[3Jt3񂝼JY|gL.+q :'4J#NJվΏU9FUP Q1Q.)?n,cy6/n>./){/9ߛNMmj>.rFQ(<3t?OYڤÊPSO젱*mCdik!]2A 'y!#F}LhpsSP ]mPZK+&#ٚ.ൖLj:Vqw5L 02pʽ{`zd&<+ #Ý3"H21EP@8,2l>R] jF]WSt3WCcEV^NhOk ߪ`M+WqwG>|uҰB6^+g+f5wۓXWj{w4 ]N.j:v9:Fy Sr6 k* k5B{ƛ^l,×' o^vɝۯ>6/=sLr }=Odu{O>aJKY__gIC{gWvzI+yBڹKHeҰ+w8do]cW% $J@6."+ǡe{Br}웇OOyP_GW' 3M"yݛX}| /kw6lhi N"|Vx[ģX[%ЃID6B_v)oUo/_n>vg' ,;j6dpMS{e#D'3wy;ƳiC6dQ{S*y#죈 jO '68w7;yȩE* n\vӏ"9Ifye(&$ot5P7欄okɩF m=O)5Pf0-1Ze02٦_@5Ơi1u7~$)%;;Ə dW9ERz8\>؄a-F'HMD=V:~0ƞSnړ xJWG.d<'T:,~1zݹĹ[Y*EaIgcFIXO>iT@k:Tj~?Q1} D+M<me 1,%؋RZZMHn"*2&@dH G ) YJIXȏ) 2ztu)X:r6BPb(ALQ~\;b 6їV9U`^@[FRNWm7-&CCPܩ#jfG6By׎ l,ʻvVAqwȮ M7&}/A ;lwT'r/s' mn*(T)v(Oů*ÏǯE+Ï:~]~Ob ?qس,ï x|!d,fj2\fl)ׄ>ڦ8{¢!5rإ,VHkxދ=B刼- @2kS2\B<. HKw}6k}c7#Fq\IƔ>c}\RQ>o(dZi#/7R\PїKjr}C-_o8/7W-yi:Mߐ%Y"h u,*,bƈ51q0"4Gj./7TLXPqb}Cŋ qF `/7'CnTs*XKFqTS޳ ߜn||^.bٻm=VIk^K+9紎aɁUxw?kknVŗٓpJYI\\qI F"r:!%x0Jc!lѽ\kp .T Ά"cJ V/(vV; Ci}<i,Q\v2lĄ&cQ~G^yiYa˟Id8Q=THt&'1ce_>x7튂䄡#sr@-xRvԢ%+ՆoFٓh`toL1qH (:1fFr'L#.U{1{3G4qN{RA bw;-д1;+9t F18Ps8ZI-}B]2s(mOVj:w(1G11@Et !<ص~HѪ=mSϧ]g`ԼuZߺRښrHց:7ߔ aR!x2ryb_"#Tb]Vp|S> h"Qk]uI{-fXJFQ7~v)œ65g|֨u(!,b>c9wz~ؠ( W.e6@~- @)/ͮ'9e(W*߆* `^ fu8hjaBB/oƋyaepM` 0_y+)n_seJڑ.dgs[N~J(iG-dOwO5NzZ.d3oi7EnM1}nO8x[Fڭ y"H(9hĈN>hKJ"ݚ'.nmH F2Ş-vXݚb#:MݞγB-i #[EL1eڭZ)mDm e=4pS-ۭqXۭ y"D$ޫ8B^D]6죖W"%Sui±^$䅋h˜`D%[] DQsSDY(ku%pNVF-+SQ>jc%H&F HJЄ]7RTC+T4%W'%x Cw'82yKݕ 4]NNgExֈ5JJYks%PTSQWbyyv=/^ m6굶#kr7e[mӽ_u/&fq{ V]xչl;mr]i<[u~^CuGЗ4W8LX;h+=}^EO_WnsAC(F{p5X\ >OZ>eѿnYH+6|6!]OJ jRNҍAdTrɵCFu'RHM (j{7HcŎ5z< lj]H|r= ; w-.NYWuqBv-!J> j{rQGW)au5.%$M I[x1KHE%5%bA uJ ¶DHӻ#FRZ͕ؔ.- 6mu%0J{V`DJ ¶$EK Jq6mc%T%- )-l+P!zVdJ ¶нK +4%"MH[] =ɿJ7%&Z򬵺g 2y֒gՕzXpQУJYMNB0h1;^av P|(> Ϫ74rN!L4^9D 3"Syb)YUPqxQ~ y?;߯o2*UMPwڡiGWz3WA(Vo y7DJ7!ⳕXR"®zM`6(}A>B!ļd:1if4́0b6ˋ'3?ɺ 2'1- b*PL-L9@*E(sIBY! vQiX [h +o$HI $yd-T B")BKCr6>G^x.tNrEnq |rB,dJ+G6[nd cVr':n>|P2ㅳ5/@ !EAHT1^>x22frzwR-y$|D=X{ڀ , @LA'b4`pP6GAC**ۺPFB(GpԂԚX00tt!)F6`B\+}}HP1t'{^g5_Osorb1\_l_SGΩ÷>wU*AmWWgSl97[?ן+w!~T]香SY2ߒon~>|`Q6 ?y|aTWCͪDϯb<_,Jd2vWLbys;5`Nadk&YeJGp,"EqMB(0zO|H|ύ0E5 7oހCs5.)[oERbnNr<~!giF4Ha]/ o0Z: E΢$hus_L[{P^w0밄WGD^?A1]T/a9I;AgΌ|h`pt:Rge; }豑n/ 8û7$JJq2H/~>.>w>|\ZZuKƻsg Fd<K]Ia\ ?|-$DYϳ gA>ϪYܳpt ckuRA VVHBʥgkEx)`;GG4QdMh`a=1L{!d 0 3̱d"-\8rxD[\+c4m+#1VU㛄E!?<p[5L8 S4`Kk5BzS"),(5}Eu!RS5sSh$t5lLIwRcD^ƕ l=lwЗ| PHTUv]hNy=oig pBzPd&od͐b:y\?ocğ0G~z0&U}h"w\Ej?20,+%.hj"D{'07f%Y`H I>٢5瑙]dE+EډNJ=YA{IAA$G&{ՂT#t{w#gʃhv.#9ӘuU0mrRXF +&֛BWN E NoYgBvpZW MBи[ yInB9&Q) )*|Ar鸔A(Ƚ愗NkWBmF>;yo¦UȦЁn V[lVtJTIKg#*"v\?x9x)<9dQ%3k,Fzjv6{cƪbk4պH"*_GEu<!yvqؘZ^?[,Ԡ=-xVBGM&:;*]C'D[m_'Soڍ#٠JM%%H'=hsm_4mQ/ [\Zh3d0q<2ZQrnr )3۾ZK(k ߍnش*̜((3ϥ6'Xk$T sK4:qFRɋr\Yp^ܰBPΩt)`9C^J w`[H#ToCqAh%6<5cwYlmՕo2Hg&77)qxS2-_)~ uW'~]Ys#7+ H>[oāpT6un;7Q$%XnnJ| $2x8i($b)R)ZֳEQ, 퐀sgbޡEZhqU䢡@kA@blFchdm= rc qؒ()t"@eH jB+)#K!1d2 ^EbD*F$YEr@,OSB@53P;?3yнBLedd$a$=d׈J^:6QI%qd3dӽp-Q(Ȧ]\v*w}oQͭ@.8jgPոzwu]07w9ARz~Ȼrrm&)X5霩8P}GJ;RF(U"}-cCIdu8ݔV庈+hrtQUP\ 8U\CNT A!1JRzP^Ifxj=:[6S`hc֊*q2 ӆ6yP!k V"AU)bBc0ۅe^F6 bXHXy(BG_jI.ث|,~ xz$23*{hnd.{PG⇅?M‡tMUrEO): (ׇ$o_{WoJ"c-=T5,O\9jdGԻcw'6{4 w.TTp]tTʣ.{0rkTN~@Xs'{6_4TG؇[$x~mk:5~ű:ctߦ~[J4ט` Ck2dyޤW'3SxC{\h(< ?!sqgBD8C#JjcзTze#udIHwleSbn@vzH㢣(_sw-aPW3 Kv+Aws+L"Z.&(_3zxWXd=\]^6,>we'߱!g9mKsvѶԃ@*q-mĪBZJF )egOeeԯZIeo(ar ?Pݞ#п포IxqĜ1m6&=m-q7іR,d{v1Z"KQOAl 8cX;t{.݁xpzm9ן} пւ`xt6%盫XX +5$7A_om t0ċ,B ;R{V'. BC=w9ׄXgRZ .ch 2ͼq5'Jp8W3FsKP,ݬ'-k_Z2Am$uf<ߘ꺁Ԇ >ak+6W6qnZ@cӹRQiIh:+\*k즒5+IsylM5tU!Cc"Js-u@kִ@i-:)&ka ! k kڱjkv_Ϸw`aq1j\/BUDft'(qHIq԰LlćDi DIƘoss2J`HX2) mnk8cJF_e$f?OEWqq{<ޖ#q &'jn?b|r9EYh@rww>CJ%s<6 x%-gK@cUb^ ',Y4s:a:txuyI^ rƞϋ+TN>L`_pkOAid"ӻ& *~*-NV^Srh:K Fey,2Lhܽ!Iqdkomw.IBbV\ r+(Iŀw VzW#x1bf]j(5JK TOWT95gSYy6<8Y$LIZ2Fd$)&'n7q]K)6[9i]ۏf.lUJ%sQFJJi҉-j0Z-FJQJi rqaC.Pqamu_`uښ? 0P()^w2Jλ䎏}Xܛi"&8O2pa%ѪρhD=%XJ [e4j T[= Ѫ{z 9LI#XgpU#J]/G7~a̚+ѝMK5WRf:{Ctm]['ʇh&{|b0~}:ӧK˧='=rx`r4>S |@i&Sbfs߱+ƥÒdi*+xsC|-Mhp/@C~]bX-La\ `d68).  }eG{JauL&90SWsʹ.9g1ښŜb8()'x#LFOTB8)V CMv<|HB2\O&Gds$Z^|i+VZp{wJ{BCoIYӎ):S㺰37m( QyCvtF6}L;#ig nc-M%m"@)QEFߖ`3^YOțaʖ-njͳKKp6cW\1K.5fpT]Bd\EX:aqJc+aqKm]2Ǵ|8=9 5KVP9c"/jfĨ#@\vJ# <\rYǹ᳗ DA?JΩ IGB#ZClMh%c_T`|Ix-ã'9Xp?㗯PArƐŌ8N:*9\NS_&8X8gTĺWdPڋTiRfq^?yOsDJtfjn:Kz5AZ2 b́ =) y/h&.#4nW0K:}u9x1AZ.q0S5 s8W9\rWfʞfRRԳ\6W"´DI+\YiJ"MQL (㻋$_4w1۾pQͶZ+q4~tjsfH`$^*\kDgEnմCph親XF6J(zDiӒѝ%ajYb B+YC;xH O@[O`Wb[HKMDXa83=f[9r5M`b# Q|m/@ I%y0."G LYq #<ڌp5f0=En`` %+zBe)Ec!d$HGڊc#yBf=&3դˌ,$1κ{_O5I%XfSڧ( 621œq@\·ibE.jS%{44 |/E_3k9<W2zĸ6_ZI if)!h5<5yLN:hJ \*x4u4S5VpiZkE{/O@g-sm@P''[I3M^!RROr:REωTHȐ(J4&ZK3 CyĤKxRFX]_߼= dOֿѦ ._#cJ~sw ru{rݝ=ʵ?-AVҐwogIkgA9T^X\ƠhD0/t'Q6=y9̛)XSŋ-Ϝϋ{nlu AIyngϺe~ԺUZe\ Bw7WöH@wx$`(g1шn&-mҦljYE]>٬VYk O'_/.roֆKX3K5MrYԞ|@6.Hj+,!cO- U=cm@?ö0HN,TEw %K=+*J(8Ϫr x™xrg0\= I9؀pqyɰlyȦgDxxqsqESE1%Bb2Dw }x b5JF?v jAfHjd:¿yI+ ?39??Ҝ_Gxw2Z%;Eer_ 9]_]-og$akcFAjs*A>ʷ"kɼ vO4C_>Nt-2\%1h$ s,6bl)'LXN3tL4ȟQ/7- >zH9\7`6`fj!,%2YDRjQS :N#qa8N9냥*2<*@BGAPLB|%cJgQq0VQR:JA{NMWt&€WKX])E57_ٌG@kQz󋔧d e$?Ps Jnx!o&Eͤ00n&7]G3-u^}_~ԔSHv}f‡6]1yp~])oI{W7ۙRm#$\ !(.@ RCq7ETfQфx:l  nxWCrT?syR5S@jBeB4[LݞīWlWj(Mtʶ$A4eP;/!#=o8_w>ji]_ D-N|yهc.n?_>|gUd89jH7QxvϥMR BhcYɢ__k7o_{QhW.Cb? 2r$r3g`> V3`~ *u&ZkDD-@ĽQjy;ZU/Amԃpq9+zjcX 8mQWKIV.3{\Dw) y\/E,An7khksۢl~z 0|9WYrwcnםbʠ$r:`p6V.>{s^իOiNW~3qÄa,}$?JVteى%bJ?o4OCcZ: _vPe"o1ȟ=TP_NSLu~~vٛQ  ȫ W]p闃'ݑwmVCd3@`̞Gkbqc;}UM D#iL 1Wcs׿_&W69@2,[M7[qwy5;%_NPs*ȣx##yoDwC$3ll0GH 4ϳ,gY>ϲ|u峛43!"YЁZ yXcpA&0IqҴEUIڗ 4͗J<)|W*ZXI|҇ґnkMeR5z*o <'"@D:PҐs@Yb,ʤ${ꛜ\zkjٗ#^HD>B}V &k8[Vk wv'9|^݉X L5z\ c/8jMcb!4f)[8{\јk4V=_?ѸJj^9-GEIy %xQ? =lĶ j ۏ1/Eb(y{Ћ\w8P@0{Y{F&$5ؼ_1H#J'l0GL3!9 ڰ@{b,%!h |Vۼ5I.hrn>}5ZXa"}V P,/| OʢC@ HuPLM7ǭTw;^4do/{ot~KeleHtW%">׋ Sb.64Vkz%$g'TgvƐ#W=|\(wg1>[=#^R:#FJF*9p2xxJӎp\ⶣwB=a'vVɎSXzԬ.HLm#^Vʨ'TQ^&Pr>JjBHUrX8"kE"A2ݯ~@)q@jj^D DB?bGqP}p.Zi!eD0E)8C@'j8fm(}Q+,{|u'1*-  L(8bQsR =!jH*jEL 햩N,-c"ZM5#h<RZ7,qQ#i@2kA$hZE4tM3sf9Zǫ=sQ jj$LVPp\ | $b R؉j[V^!UBї!T=7 aR^>ZGq+ۓp3{j%<Za>ÈC&vT9FDy ꅵT;;m` :1fCgJk"}]^*jo$w$/:O!c *:,Cܲ|⢆(oa 73Po=*K}{<8y@ۆXݹs(,p+ CnP3nΗ>^]}ZM`m|W,b\ᦕ(yӇ s~ݦ4..W_8p &6hjLly͌.`12 t戴X{2>zQY7;ן}Ҁ)&8)G:$Q#EGdmq)A +mHr9iE=RL҄?UWwM7Srjfm,M+ٌdaԖŕ"Etoe1 [(j_4)ŌRk VLQ2MCVF\7OFg q6VމLKڍ4(]r6DCS\\_ؼFl\ÇԷgCD1922$vǁcgt\5]أݹ*g元3N9I P9h 6wEy;}+:.z6.[M;|qo̜spsك-!p[`HBXyۣVۅGUXjQ_ %dyڅfyZ= k;WTkm$7E[Y); l6rkc{eyf/~HI[d<3ZͪUbh"^ nzd ƒʄ4** \0" wXz"0N"ކdpwAt/ݶx{Jޤ m\k9*hU$ ӫ( 1T)QP31nɼ55*쪯HdݱzU؈$kC 4 yt?pa#ɴNz':M> Byi4&W0TZǡ7a-ƞ1BrŚ%m)Nb~#Y vؠf'LaZ-D$"fU#y+*-Bb#\pEi2.b|DY UؿZh~| 1`V/ zɔ.cȡ`s IkM TUЄycbDt&+KZ7f d vn; |J4GB&#X\ߠ i^Z+p*㥮,JhƉ /U+)2%VUJQh+^+ Q|iɍ#1XC1J|I=חMկbIQ&SԸn2{|d3Q+e(ϔF(̆ebub靄di Fp-W_'Gi!2TIUw"3^xvxMfzR&P^81@UoV4 VFJ0@Vzb>W=JpQ7Q=R$51)ƪ u$&Q /[#څBhB S |DY?Q@dRη!< %80R,Iz Oqɷ7}WyҪ9)֟YFEo_ f^,9] ?nvWq/Ⱦ" .`? @_idI8*3l`J:c%MTdi,oph}LfV.|Lu'q6Z L/cm`ZjF!try[_iWBk)պ(! kxva6\OK>JeJFn(`oB\Ӻ''>u4;Lo=NY׼5{ 5:j!2y\}ۯ|bfyY?Iw1^< 9w<ҶZo&AfFs*C3rh]CWif+EZ2$YQy_d%'A`+$cd BŐTEs!r9ufNX`NC23FxML|Y9#FT1TNxe:B!yW,VQZXyOi EɸGjsu·$G򓁝@u&m]+%"v,I.#n tJ{1Hemm"nEh-բZnh:[$a2qZh7q Gi27ep D#(0_R7 :_ڤסiQ?ܙ,dS0LXb6 9IN]WXEn%߮pm"%u\]+k"aoEӭEb^S8{%Kv7܋iYȚ=R{>r iB_oMs{Gq=}#* l8X>;?#Ok}.aZG] k v&$IV\3N Hf-c`HQW:Yf2cπIѓMz,SF"F2TvO[b-ŏ7#ջN"uiDu $-cERܛ!1{& UH>EnjAN}(6owyο?&wxVx! |:Rʩb^&}Pr.i>e]sVtO~n,"勹я٘*Hvp3;ŃZ5yMwgjCebUy!|LEif~9#"5kI3h"xTPQG 21HBJ5RfG{͢cz įۛvf P!4R_kVs&jBbu;d˾e ?̓Zϴ+nZ$˔hH՚V8h|% @HL~"AA Ӈ q7W?t0Eqx ±~i_?.wg;?N[ٚ1UZqZy;i;ud϶*&77]+hJx }Ҍ?򴆨?/ C(x=̀:o\F&}8B(m;A:Jֲu#E _("fvAn/gz )%zݺfY{ϿޞiIEgʦ6ٯ0$QJp֥.C1f0]J& +\ 8^̀ᭀ̢>.2s6ƛ5.ˋ0yqxX؞=4f($%S$l0\ffgsgf=w>^O$|v%ד >~8r?()R ;@vD8mL6DY.$JǢ?X`N٦CH3> 1CO4C@֎Ge1-&YVP&Mm>EZBuFT1`E?-e7}ښ |i{ݤa th T d/i9M2&vٿc*(W!уJZ(CH䍨LCl2$ BkȜ`򰲬G| 9z u8N $:nmbf'rF<_Wvk򆋫 y2CZt%itӌyG5RާU5ܼ;{ǶMċz߼<5ևOןZM<:OFiu(NJ s {yd;^Kkj_R ])* '㬾6Wᚃ, ;э4̨o;a#nPqsuG$r>>i@$;;mq2*\?ىaS4J8YtkeԮKNscФVSl-oi@!;^"">ɿv/HŦc&YY_x=>_n9̦Y4>JǡK*80P{{/ruEW.'㛡V<2ܛPT&} LA#xCmfh :Z|v&sd_m:+M8-1Xxz9փZQ*%>@l@6$jo&25a:$1L|$ٓ\dOr=Eݓg"F**.A*)dbڥ ܹJXax@#8ĶmG7@֕| d3dͯ~{ߝxwޕ57[n$:BO ǎw:_fC#O۱}EJ,^-zpKb2$y~XɻԅiÆ;-k}ؑ]2`Ǎ X7iVA('lDc?E]CPOnrd IecVۺPZTD/ғ`@F!e ޖԪ2p7τVܘ0V:w}ߌ>٭?_qJmwq%$oo;8Gbpt KDiH")E('-(ҩG *b9\MiH?7WW 5}xt%/%:<!R1NJ2 5 WPM 4 SoVmm܃;A>uJsПir8]N&V4G#x9c jhP@.^,:fho0#e| FzJU+t>sr8De,c6,[#iV=௓ >OV?2V#|3dneTBid|tQlX}\%+RSL(fυ-t2LCꪸƪևU! o Ylr Ph E]1ZU)RJ\ $߬ķx2^ JU`bɬvY:mqIp6GB2F#ԫ҇ m3/DjMp/"2פ ԰y7vZyznhDf9Wg/#-~xnhK{JZJ@!$S:k;7yp礏 ୗ\g$Ǵ:vla>MTwZP-ՁT%'2'"Fl9Fc oO߾*̞YhJURooM&g|9IA;fSUov~u;vWW_%\wE 2'$X%JqT+\Rt8%1ZRޗLJnC5\ٌiTcI(7Xa*-'$ۃuZ :dP`gD-Ղj B~{TڞҊm{EUv`}O{PVp> GYz_hw? /#m\[cN_߭T{d+ӄޤCx/Gf&$޻F0wnm>LKwVS>^5V.V׍EL)f%nFb#:mD];@ZO4Wu!!pm%S<{sOyvkA}Gv(in]H\Dwd*3R j^]w!F.zB S(^ȃ#j酓EIiK>O/7NsUuBڂhܙTAy@Xϥ1jBf%0iZ+Bw67Qg@YaI )_pEj'w\(LʡR!m(HT״:]$ Q}$9F&yd21Rg:T_*R+<}[`ǘa;P:8pNcuCTO 0ߍ9j0'ip)i+:K/"p`(DG-QMō*'nddJ4tPВr`)%L@-EQ|USq{А(@nkYXE|rAfKN!~F6O>E{sPd0Cpm.36-AoU}z|u~[H{G6=HPx6me|3lUkY -lFEJl7ʖ}aO?.㏓Džp%CfɦO+ԯȝ)L[jE.ŖݺR+ pFt97?׹¥8 SI?Peтb.L\y4W? avbU~:֮O]*R9n)8bgܶ!_6 tR?ΚW_*BJ9N3uUDob:+Ű36@Tşգwvv4+նܒхE\5CaH9JRz;v SoB7D_]T U% #@մUp֢( fzoWm$vҽpBfGe^#5a̫œ38h^P-K_=K'$1Ibv.f&Bq";'kNC?U@cȉ ǥ)"y@tHd7oVM۾|G>2oq9JWqEһ}$G>i *\Т8S6?pkIjRw ۣplw# aח`h" ^L6{?_gD~]6:+Ƽ6ͻ2ͣyNݍN,}j5x}0.Ħ|735YTnQ?_ӕv~s05jQ gZ6:{j"Մ&otNY0BHCnP 0m.fT?@{Wv!/8jMj7m6hA@^b^ XSU CutQRx`?&aS+ Zy5:5 UrpT.u:Պݖ2ߞRޠJOs.J]]>#D'q5R0);3TTOɾ=\KڧFD3I MSm!gMz ''ژmO-v[}jy >١r sxykon*/'!^_lلd㯝ޠ9i퉋ZMذU 4g>KhfѥXf,a,9} XRs/Jޜk.NSf_'Js:tb6v`ëSSJS ξgvL.=ނ_OS)CP0)p<@ed#xê͘R&j鸗kR/לrm0y!dgf*+ʞrN1BH-e,]٤6R+㵄z87w~?Mܛ$7Ipo[?FeV,A{ Z*(r<>FEѷ}tdOfĤ ы+ObrZVL[8bM V́R{-(Ng{zv"6 )߮8jڎb.uGNp֊p2҂b΅aCS,3*HPj)dg^('=V?oXNzfJ% WRPʒtD*C ( ; 8k3˄HpAȾĚMԻT Y+ HQ1a^v~UpEf6E.+Anltqi0}޹JmIsA[Niդ2Sɵ,~ů Fn(ʓ8}R2c,v;;j6JcmڔvQjTm& ҧep,$/$8 #xyvOX㨿uXQ.|<[.}el]Zj L[ޯ@ĥԒ7+ODkEB^ىegdn 3bpy^u>ɜoߦ|}̺bMC,r5X,;σ- ^S R%} -'Ɉ.޳67rWX%CTWueoRWg')qN֊2I0$1PGJFw6vv0afp#qL󙿿Z}{LHzz7]}AMK-D w^)b#J*I&7jA(2h57p+P,. <òoJVPO_|sВe%Q(\Si0Hcc(&(oq$$  JlI^jʌBؼL=N _=0u@8;>5|B>)S6.9'<XdNa{pt.8jNق2X'ޱiԜ뗮LVOOXH"=ys'lQk\d1>+/l)(~ټg+6٩C0h=bA GQK&߀#.Ax1 svr]`xN[#i%N!|JW&n n6oz7^YCJY)/׬DkS_2+4H@rwBY%߹,1R3))J1; pSK(R/9{ew~ц@DRʆ$NK¼jjX5\z3R{UƖRR< g~GP+ѼQYBcM,QHZ:[;&&cY;:(JX-HP1Uz\YYR]݊odWgStQH=- [Nl$c2ؔ3dH"9@o$?[R3$4 iV&Nyp5N+}YJz8Hڰvp'-h]r.%=O t:؊nFx" s?0 YjGi"ش Dgs0u-C i!>1GvԠ$wFyIpڑVܨl+L0:™"ؘn\$FttFG7[qؖegj'6l Re:#{D5x[etgIT !1<ny A=ï>^9iiOzӾZ9歘Ռ钞wY>=8 UܥGA.c# WNrŽK !͢v zLKydqPS%qx"qJDc:OhEAQlrcɋtӱ$,|  o L?|WQ-Us]9W:xς ~(ΑPDоkBN/{T6VÁ3'+%[f2/qqVXsWΊ.Q+YXc} jbmwypC-=K)+ͻ\;"vkcd N{m~sf 蹮Q3_>y63_Uy+cWvu!iβD 7UpTQIcOu`azq c>sM K\h@ j9՜cԬ]-,,,D`e4P N uޜ!PEDR$4wLT"PM;"b"^#1I%%B2x& F9WI# HIk5A)sY0DR,+XVKgDJymKAFY`.9rDHqi߷ *a#0+ nY=**55\DLNY>JHSSn`tpӎR:tHD1 z 'NK;"B 'yi:*znזW룰wٮ*ע@j(Z߈$rRIg1PB(JF[A wh-BR8\++i BLbJ^'Uª33jeDh{= ;h0,!my(Bl@RA0?U85ua"T+b,et> Ƀ-jNKm.7#a"А F`(*mS) u>@UL=EW8+PSPxO;Cwm];e, }uO5χY"U,IsA^ &!'nMt?wlL @s-ڞD|qO|°۸T`JB{ 4d:f 䫹)ﯹx8-X0 ` vXnAoцIuAi2[E)}dO_% Ѐlx7=7|Y\ih`8l`NY/D[xz˒vLi,b@mU;WDFHy]"[_/! 3I?,ǘ|*#Ҹ")trzꢯ݋&:J}:RvRYp渫KLI>pٗ@툾D;VY m<)c$q.ht́opP% /~Aj+T`ehgy kwl3坣Z?oߍfr;YZb<},}{< AQU=c]doA%,-|w.m)W*eKHq/m0$-2I #i(S:N B`KhU( %DzT-K0pɻa$֚`1FD`NKVr:qTá &D %00uJ)+1*(DF-Yb5OKQJ̬Vch5TJ9SVN: ,,)+7JWDO~ݙ@/ hK/pw/2Ojk6=0OkMGL:^X(aGӇ?E+LN k)ÏoGnoG {?~jZwNǿ={]ߊiYmk 6]vAK^ī"͚;}Y)lX bUX˂QॴIn|ϛW~+μ5B2I=A~A%oJ&w);bwhd`R~7}S0$sPJ>]!8|tS]>h- *3B(OKx8U=*6$e4LYHxO{E A(m¡8b M%d{YH`DB;/j8FRԔNuSL=w޼pRzk^8xBpguФc|khtghćv;O/;IC&jjXWhVWq̊\ \( `3N gqɆu:g@-^̀JsW3NP-8~Ԡ N i1f`1 F"x%'HE51 K\b![ ZY*h9pNd T}«7" :*2W|*sH@ШC(uؔA 76~y}*g }ԭHgAMīoPZTNm*x)'3d[j.pp=S% 2Rer0~?M 55J)WUoOXո7{Lf0Mߎ4]i܆-']Tc6 b2o^hT]c3:MU8O%=>!x)y8s{Ұ몽/|Ly5V.3;߸,S=E3dGrXjĉ] #OKMRퟖB!L+MIնtӈhcIeY"s b`& Q҂*:sȒsF"*^!. 4=iI#LaCƔ+ :V s9& hrHd$#݁*lJ% i3EDӏ*1qFpUU WLn,6[ڽZ )௶eٯg8 ,V7DT#AB1]XOa)TvBj,`G@XR*c1c{KaJ'C!1c*K<3+0Ngw Lkz ?; VN=-PT;"w'e)~Qÿ64sЊS4FkD X`>I$ /2*2qZc\IZE 4Gζͽ3uۢ h/ZO|六uRPu|>.˾_|T.;c]!\x7vͻ QNE=^.VΙˁ`\ &  Hvܽf&n n67{Gїli ,`;2Iv{,ɞ]JRb3Jv-[U/`Ef9BP, J;k:v@@R%JCJNX&5j5qʹ|:*]u1n8g#{sZudy6sH@Mi?"\@Lquo{͈т3jwuNA] ]5=g@aTC9$)(Xi\CD6ŎHoaAfu3⻦F8b[̦X ؇w6Ƙ:d 9:яJQYaI2@;Ho l6fwL D d:iܽZuʥThl~HݞwT|\',.rA>y~('l_o }CՇ7l-'A_k$ c8u3p>= c$!]X#/kWuT_ή[p)E(POL  CD/< a8#+WkJ_߼TNYU{]}L^Qյ[uA mį-n6PtR6YFxZDkBO+?Xnz(DiI XFPZQԱ9REi4a(🆻ޚtMm.n?mqϿtfl,/qCql5q@sgU9@+8;0J|mUH#;D{G^}&NC8|sN@B \ڏw)dlUJ╽QF iހ*Lxa}FJzwT(\ $i`nSib/L`m:XP49ޭvİ|zI~/ `9|PA\*z{ۆ/<#}|:RM=V=f*+SԬ )bFDڰ_.tߤ,M\8$o$r9ㄇlo#y^ jD Bpdi)vL{&66߻PmvU7cRt?>=7gAt{?35y[#tJfZ3dI3v\<U/FjͣBzYq˯trٗTu͡:%5z6NCp]fdI~5Wܫ3uGeCh7oCz!g)uAu g$E(+el Jr%iZ_DmԒ1GcXs 11%MVIC1$HJiF&]ƀ1{򬍘M6Q9:-  \ͺ} ыR!jzn_ܴPp7UNP;dWY:.u$S(ocv8H.h;R-j jN"qy H[XJZsyhؾ8%pV|(]VIfcjv MWaJ:ƈp!x:4ZLij`;-s^hKn XvE ;$*.vRa"WV/ /.\SR}ʺ80E&kBcvY?jR'#Kd"0̻3xXsoqϢj^>W pW£(jdr #( BKWľihc%M)dq >ZwI C=[b4HL|П|}!NeA[Fmz+pF9s#Ğ Ty;jH?/EG٦79ffSDyQ@7˿hcz篷q^]4cPow !exsyxV×_puTa[Դz!/7?r}Eh$ Mt2yKBm [zߗ1b|u:u>ak`=,(.|aaJ%!Gҕb˔ռ+;7ʬ!]DXhU ZrfD)*UX{qUx6|{ߔL3Bwl_?vVͻ[ݏ",O  ef@yƜ@fMG>˪:>AxiE^RP Gh<`lRN:CD{$XE+҃(pA%(&MHȐs.*S/6 k,1.ixޫ:h]QK ^EIaU)_)FDK`+JhU^w{74, C6NQ%8 ^IThkt>Ԑ2** ӫLF?0d6|w$7M3I- =݆FAWʠ8p RTӔ=U-vAЖ@{/_bÇOݰfפINh?Fg܍ ߑlqo}zraVkA>u,L8lZ0# }yx)^߽> ͱPE)IMn=×޿X^F,Γh/Nw;F()hED. gҬBwrg4W:͍~_LQL{u3V/V 8"GDޕ-~1i "9CuC?(ٰPj| {]גE9~Lgɪ9hkmHZuf,fyZ<;2G2X9v\rIhg=NUA9h㪽:PԑTB:;d2/S.vѹ;ޕ!Z)xuWeGfB+ e6/SǛ8ޔ,7q " F99a 0z3.WR}BOW/jj|}T)Rzk&v *oJxe-V1.ő6QC le 2Anצ6#GƌT*^Iј[ `7^f.EcJv]ykKbj{xr14֞rfK;~~k<,@7Q ޟL{ 0ig3Ky5ehETJ,(E6%dcՃ<;ytu5>-,>M}^J5GqiJdmrY<ؙ Dӈ6SiGXF^BMȍ(&. MHiW`ʺ:)7No;)vxӷ~T M'!E!tPu9uaBzCe=HC|uT닚Z_c Z=C>I5  Oj)HV ^Yi2yLV*yiP9 x&'%߻|ma H9fk*r$chpF81m~zSLH~ PF3XmۋEw_w߿u$dt*.""Xu1n%ٸX?1k/fy mB6Xec1> T#*~Gb>^yn3WUKI)ywg Z9(T\(j.#+KwNjgS7F9g.2sJWLPoUP=Kn_#ipSM7eAH / KN:{MR23~ %AGp>@;>h}QӑFwKaP$@[-~{֜~wQ#TRXկqnDBc6%AABcp ģE+$J!Npm߭m:5#cjPypNJ<(UY$ĨbrA<.ν?̔\C\S5ɘT,P4JFcm(6ȀTPZSu ?TmFԜd(]?y7Z pa\IZ NɆZR|"kF3`}ʲBI2ctQvxyH1vPh{(vv:Py' %8:pAt8plLԅM?{WIY$ l/`k6yYC -#wY%)+JEVMA# sT;쌢^JLU ޱp)"]ox.[] b_PA4tB~4 b0(qiV@5L(Tß q;C{sCCPwb~1;W fQ1[j+Q˒ĕ@K4P#Dli m~9lf`u Fň(hף+p;=$&hCm{GyP3<Jv9ҏ}'(g=,(X6w9 $%V%uˊ4h}\>릠;A,P!p2@ѝ ,P%ɈQq}9"d8O0?"e@Ý׫p'Y;Qdž;K2 w5(tpv*kO–L-"Zp&8fo84e1i ?moo=pߺyN P'KPٮ,szl( 3Xgd&O1"|!'~==ɿU9Iߢ$˒?)Md_bS 1nk'FIO;FtB$$ZsBrO#^à}+V92J DWe9HEM.|n#eYXRK}[MS+3p0v|pMԍ=ڧ6ֆ=~ymnzGp:gSQ*1+J9?}Ł?*PMNg~S2>IQ^/PiJp6|2)%vFjH@Ľ[ܔ/~z>mb״gbWp'2Rœ/cŵpWƬ貲dXZjDE605m%IUP'5wlfK%G/Aj($LHYqf^HKT:n ՍBG̑ J"rmt&m(JPxH.{!oiwFȧيDbk5`(6`BzӰdK ϯ[>Tz҉:Zgw'b49:;ށوHh2P))~H`J k#S狦92T9A0K9uxl98l3RYv eBzvv[֨wv裷5MsJJs-SkoFX{ j<$h$ף+q;=!$iT 3/U2{azVgaX #i=c5$hI@e>(è`$?qDˠS;zɨ)A%$/Uǝ,Q|SA|nAAXUa.=KBC!l(D 5Vh@^G :>/:ޯ+KL9pZH,-3`HQJ.ɧSv"kcw*&F;H1A|:C 4sojW^Zl ]e!Y)縌u3A\-fo*͉+@Fv7D&$3@6ޯ?:߉s#Q1+ s{{nQiI@fE^ð^ GAf+0|+JW;< )V>Û l C͍b-$R7Q|:Bu郑1ᵚ HG;X>A+ÝHPY!|ĸݫAyD| =`*D-I1:>jG!7<@N.>)keЦh '>ؕ]Xm0Z#vG A^7tH h vV YjT.x{өchh-9 m{eno6H#q?>9FvʳOn q (c(l(Փ^yjTɢB >T9R:a&(%~Gu(ap Γ2&Ck(aKczoεը9!w$ܢ0>G-TQ@pm!@CEQK+]I[wᤣtio`N9L7e+~&aK`U)>pjݿ iyeXa}^g;m^!*+@JjW)KKNR VF uN-1UyDYi_Ђ'5;ɕzˇEʑ-DP L9h#y <7ʢ4YZ1B˲m4lLSҔ`dMqDn A.+덲THʎscfGDł[zEÀYL֚# P#qX~zz 48]%둤Q QG,)xI(?ĉӷW$D zPlbǩdE.o\l_Mu}YqwFXB]W@| >?g(U$j!`&HdY^|#]Z)Oİc %~[gaSP[S0DV`[B9,V߮<+wfL!*6rgJOwP  mAkZ(Y(BHб%-vFD{6 qknt%t뷓NtϜϑ,x7B yh\*"o*\nﮚ<>_\ۋጺb5 z#=Vn&զ 2l˰)/ۛ}H+XJsT .ⲶUR^Vޙ Q*q܉ns'ݮa^!;SacsejȻF!?Vۺ ԶGZRǍZjC *W2[)C~]) ծ=nlHeGgh'B[Ax +΃ [Mz3 GJȽ();qJW| ǷJ2<(H(!]x ̯Zw0ӽ$<>ȕdٷohDO.y8 \J\@"'|Jz.( :Yɲ.?۝qacw#F;3WWKjNIэ? !s҉ud >0㨉$%tlgIn!b-F5eRbY b.uI7 1HEQ{\'ER!ʇ+UY@FR J3bBm q35yRdt}{Re IGz>ݜ9COs}p՗5=j|.n?]\dK>a'vZj1o CNF܅:W>>9ey09lƾ`B܏he=W7yU7ƌE^Lne WF_$ѮW7ʟ~f#eV)tp P4C{"-- RUI+ĄTFȒcU<ݹepzk^.er!l72EٽSsØ#=c2о7X8x{dbT.yW>M1Ct"zUҥp_pw9+V]b|IW gb/_59UYKZJp" FBm*Nj^cCm\4q͠EɣiQ:iG{>;M(P U.ZP{Hr:=trwjjE3'UdkIiIi'?}T Ix:gⴡ#\8*ҙe#k.\pT6u(.'\80(O*83/2X`.el[6Sj eiMpʔB0uyU; We5Neɝ( *w򃗲7jyHm@3\ZkH<͠yI(VZ9ᴲ߅ٛ*YZs#ÂԺ6O7-\ڛǫ~!Z C4O K[Ro}_oZ֦VRZ nK"Pxg~t{,P`zVrȓZ;jEN*=_-J;8)xTqt.GMPKմOrG՟;y{cNp8y;SiYk*^S pyx>Z9&1RlY jk+C/mVr,9VQ$9w$ "a"Jl,cA]]B#Ȫb;-TjMyq4pkmVEٳA,x3M&Ξ ryI`2%d }-n[V_$ۚ f2V]UX*  AxN*u&(X7s G%z1]Xp|OO[d۩W, YP)wIC$AR;kJ9sz5$W^,m,Op(X0,r©@ldj5Yv㟱< E{hVgP/YֈgoKUphʑ af&gE` sA;ԁՅR9˽i5uhyR5濌MW1x?ɜMLUSe*I[~5}vn WylP^Kg0WwEia@n|uiEk\[Aޣ@/u[=i=hD-n}W\N ,.[r'=% u㾇n Ehp"4aZ\j)¤[Qtm`hF$!1 w^sE "&xp)Od /Qˤ61|fsp 4'xn8cK3FBK3<-`a#ް]G+'I2kgv\,Z ;P(u\2A .]eQa`R#w\jx@]yJ3jUf5wӸ Ҙ숃maKK2 4#Sz[2ʄbZ+X.:H'suPn 0މ-m_[㬹A=ߧO+W R蠿}ҹE; w3qZpFvwЪ4T\ sG̸’VҵvVLj/+ayOg-ӿ侅mI]цxڰA_0 т^dLQie pGw B;3M gS?yï^^ ]޺$;Aj?b=mjٲ-Mo>Wvn*z2:`l"Q̘=vo`:ӈJ5=/)ͪ$k] hG!ʨ) A(弌-~bkHYs{}36&V]|gwbq- 1xx7)?‚MpiUC|k3'6wq}:&ѯO?~=|)nu߁#ə -I}W@dFbvm,m4z(cN6Bl:|c& FÅwZK˜5O\2){| ڐ $We+mYƸNI.(&D6S-gFa{vpQ#9A.TVݟtyij, QIQ rf_u!x }Zs׊9) OgA 7nn&Q)K+..~~~O5V@'KgcŸ1,4oѪH_#ZrvW"r;!ݫ19׷vZu@ҽZU7-x9^HmXw-;|&wc{6$gfK7lF6^RmFаSM y%8OHV 4_d6Lc)uKYd nEz$qhÒR |}_er WB[Mu%5 v{#rQTiKOxв#޶gM+ak 2?l`Cͼ/b׼Vjioc HbN#Q?yz;%.@&71Yr]cߢ*My]ν(8+\m-Hi$Kry$7(!էYztC-hzS\Z\&d/2ٯ8jhCo= bd d˨fjh)c`2 29B(.Mj椼(B=HĵPBqW<'m}F蕙*BHYle;5qԯHKxuMR<6!_S_;'afoBm&! Eԅd4fh|,x979kkf #Idz,~آv<,}/LH(V$}CDC:{3Or,w_a3 T@W+Cв!3OϕAod/NoD0c"Y<+4;O5"~ckv:'W־0Btq=|6܂f{NvʐkwZE[ `.1$.apaX:}|5 w[7+5eAbК4U_s #x·H8bKkvM{[R{#|J%ERgtJgS(o: Eޯϰ6LQz )˒B oMiPLDE yxT}:7ٸvv:MjB;څY^*FS?'Z0<62`ua K_u.0ICfR9r~jȸ~(]4`e8ƫlQdE 4P$kYj32@/"ТRw>F7%ٚn[Hӣn=F2lo Q|)orvd󅑅+\s@E2Bi3\ǠS@m7UK gr/YSckc-AӣvҨҗ*X~ǻNyE'o>UC|M+{]\3ѯO?~=&oK|w;7}:O!~;~׿3u<;WK̾#h#Jp]~X&NjAq\3RZƞz32Qd'OL'ҮeƘ7I, 7pGfٚ YY:/q2uֳKS0M>AD&蒊n5|h+֘eӲ;AcEŃ͘R si筴CxHuY:AsnNxA8`A@tJQ tOy)[!)AطAQZ-`p+G|tN"I\BxYGoJi*g-J;Sk$A[B0irǕ7VuwG+;e`ɝ0m@E@xgW+zGTZKd.;2[0ѩ#LpiՊ/GA+ZopRWSGRѫ 0̂DX03[y:9BRb9޵ZJVn JuQsc5TF$2A &bJa kB^::^ǠE^{$jU޵jܦ*Y0Jn2 4Zm6*哏ػқ()7ɒ JŞϫǢ\8F&ikO݁nPMvH۫X:vVm*}I<ܐ$c{|9y`2,!Wѥ#v>Wy'O-'L'Iފa&ۛ0_O棟q%9N uM?OFMGs=zI2ƛngG-VvUfU Ph> ,}WWV8zv;s='jӱhP<#'3Y F^$c"tv%߇$x218'V({aUJ;T"F 7/{ZЈagOt^$':F}9k]AuM/ ak~6#Ek2QCXĢ-31pq|Ikυ>[Yr Tlf9ˋ֓uXT&e֓'ֱU9bKK[h";N j+V#m}|Mn~֠I$1ab2qf4LvoB6 p\!v(z'f!${zH$ g JbċHÞժJ97w bf."#{|q9?:Ac:?`kk{3ʘoʧ0ӯ*:8[jEjm ;T<",@s}]|(ݣwpo/B>{!{4b;)0HrxUj͎:R 5X>hϫدfb2:Vꇥ>  jmܕ{W 4dbgZ"',a "AI$0H6{ƈGf_G[[ňnWzu= n,+ Gu  2^$4|sݢEpl=" k=pێ[p };Bf-{C OhW$-(AEoF ʃbxu 5ɴr^Z+zvMƙG 6Ps8j !vO&̓c8VKWqq|F f[҈Hk c&!aoIhtaj}=!ÎV?=i@'}5Qv >rdj}%\_aV&/iX6ģަy_2`4B_ga0<ڐLo-+Q6>9/q|5;٫}FiHzu &hR0E湿v= 9&Ed݂,ji[sdC m`+6Q}\6 deSCPaTD%O룶[}<VHe2jzϰ6sqauaݒ ~{ 0/y?.a-E1]Rhc۽b-#TĶ R JJcWkԌ )H7?}t.E֭ 8xN6dnMcP&2rpkzHڮMC(BT"yP#Ee@-=x/$)OJ|ILB(2Nbցfb6=X}XP^e*XltpgޚIJjL@*AĪe_jh-NPZF B KNҋRyrVQ9+P a~ZC^pݻ,*u{ugOgk,N7+3h%sFG*P~뼇5xqV?>~T휾.u^rd9+9~ŭL.T':9ɖ;' 3US[\ոYE&+zwHx;V]h5 Zv߼zwE! -ylz| P< uuCy@&U<1m9o9P ZZMrNye߼rP҄ prr`8 nګz+b॔a'`HIa%˕ uK1a+aRcp*{c>o*|0/:ƺ=:ā/OWVp_g/[iև9@v|:j?idv~=Aq?\sf|Ɉ|;K{t;;Ϧ4x3{p=.a&.n?O篯AF+}{^/n;7ՙkj~Z;.H(+6TZʭ~ΩzQ-SHQZKDž휀 rolFZP~w6ZݿsZx('!W{=ޱ{B=z+a֨dK<]SޙE7۱-~9jp>~3ݙHY uF`'1^AЦ3{> .0' $冑iZamR60&Ķ#}h4tm7F008+2O1*<&NΧCyLLYHRL7g*ev:6c Gӂ&|ԿcgRԯ46ϞTN kkG]F' ae1 2&&uՒlXDPe)CG-rC>rw[Qۏo}OdVU-;{G& WF/JDBJZElɗ*< T$HzMl$: *J>3RY:>AѶ6Y^(Q!j5Р/'9><_+>&eYgYeYxVO&dBBqgGE< ebH낄RzK2|>oOxn;|cv& BrL*:5/! ^?/<5~s_^>--siߝ3dZ=B:.ogW'ߦ跓4%C, gmt0Z'd+] ɤ,!Gɩ6ɚ"9iT |ʇ388ZQ TN:_$I;4%Ȫ91Rl۴T]f@L,>G emocs HyW9.Ѱ'C+Tt|]NYh* H% d1CU$ {cZIa4&C,v~b?S`"+ x,]Ɣd`4kM")"C]֛6A2RՕeInk6b.Ϻmy(QXK 5Nvy潱5v)+e.ϙz,''FȗD%𮼨̽9=AVLPG߆ˠ&.O;z6L@|hRd}!';5 'Z#ݯm_;NX"ΠaR XL|nyGhWgDa,]>ςld1Ekdzj20aeeO\ޔhj8n1p@S_s&c`C256_]'C%#l0tk0VȈ7Fڕ=LV7oav-Gؔ*Y?{e:|0B :"iFטNNKR#j=a֖c\j{Zk_]Mdaƫ%ٗ-1~1$Cݖ1bMk7B97Gjr2H +}mWڼBnOj~bVRRxcRRz@+wIkyzh.|Gd놗tu{&~owW2d6^^iD~te57[f^~2gtZ;`vz4(opaф1Zcg)FvIu .@o2AwH&S7"d@]u܈B+"98%]ۄPk*e`fڧ>{5Pb5㳃FDYw=4l| n"$0}eQ+5_%`lmH;Jt,vW[5aofm;h^B5Tfgp ݃m;m'x Y 8@=g N@ Hֈ{bf3+uaG>e}*%Ge)3w{Mmߏ[N*^>7\)qI5[ʘtr~YO:']˱@.(ҕPi* a|оtA"Z'o n-"ܳ׼*߉{7x}TO'MG]onS՚{١y/Fs_f=Gϩv>΋7ߘaMhBRRgE*1a$=ƣ4bm4 OʐA0A"b*G+(Â1HTv P* P2$)G$$Y!ʳ1I;;|u}OJ Fr"!@FKHi`?Fpbؗ|J uW9g'3zI3Q%{JibSS`^/V/Mh-֙,8riΧb(bInpVƕ'M:P./9%P ĄLgS]a,A-BY$n+Nc-r;i$XڳfBMKX?OusNN8bVP*Rlpb! ȊRzW\3<8?Y ) RUB.,;%= *,0Fq.F5*$ (L##eIF#q]\ep(o$r@@byqMSL,9oR$ 55<|e%lYI;zb9pTɆ,q0С$OcS.(WEɬ1(P+R!# AHLE(z p, WL. !t]0 "T҈w1 b8Uف.U"ɶ\kNZYEJ$aAFLX Բz\N'I]jJ5Z{TԚ) ΂鶅\K+(]/ശ}c $9*]m,5 iNm}t=8zW 텄eAׇK*dx+sj0ci48rU8^2hU@CEExC )ȲՀR{#[ѹ<Jc] |E E1ͧm eILX*!a'1;%':R.N,!5CavP\ : U$r?r='pFo3À[Vp`UT UQw'8%^E)NN6J lJ!YjMk p:%XʙW  NMW@"HѥPʅlJ"1N,8PvO@BV7Ȓ C6V٘cpc[RTV5i\ε25Ury)aVDxUD;[Si pvt`yRLUWcoVѓMI'@MaucdE!@xۑ\ClQ-Wk,J EW(GtG"̝`TSJ+l`XT@ѻ|ʢ2 KNK xpd ԝ6FB֔:s!#g?(2Td`/99"O('P@YZn6h!22*@.JcEAhg8̺z^ }WKG2Ua99\ q*5;8Lv$b+T6E8b8I3w4-)(2>88UG,iӥZuj|p.ZfhxĻh`8g^2 ;oYqՇ:A K7jpls{vǚDUFdo7Mo﹃i凶mdjI)X|>X?0έv.viWo*8(\{G1jz[4; ݱtVKj8s}=Q,zRg_'jCv:+ 3X_QcPޣ(u3hG*0Fȝӻ{0!ĶFUx?%\~wg1fȟn?]tGſ;:ϳcX5No>-~,>'/W"K޵7kpޭp뗬u?-[FBXkW݊EQٮE3(zd ~ww|m֛/j1vU#!4BZc=l{r}Z]|{NysYh9?;8;}>?,zu8F( +"@f0|:S, ~SX*Ov z0({y63p5 n-/ods6g 7889kW/r<'פ֝nc\ͧ+~*mx6^.nDl6|9tN_׈7;6ԃzoJAFXGj>R,v4%UK?DOg_F6-jc痿.ۇ^ag>)W^^7ڮ4y~N}Db7O^#ȿ~!ԎfXʹylUߎKkRKコܮ^VV4 'q'Suk2I(7ӵ$ d$w2T<43-u$3E{?N&pfrvjj_(ߩ+yK#` lPI[ƒh:3[A#X31b^0Z~ѷCgzoSyH{^XE/W}vo$XR`Adf@@;lkgzu0{Q9L!bN*z-OݐbdK;i/Z8Fu8Z?| 3|ߴ->pzpzp._1@g _-={7W-y\ڋ}b99\ʏw܆Bi(C0Kԙ&pzfK%;y?u'rus KZpd ^h(:'I~<<ƭ'ޜ-fV 4`JF>ivxKY,}56"tzZbN<:;GX !zfQ뵤CM8zLPע.)-H{+?۹\::L-{Vh\Clbp3ZE_'sç9|D,=mN#6ăαBll vZkG^λOyC뺣h̙mhLϺ^F`;h8iјQL%W{X7O !\t׎WRoӗݞ1$=ZCa=)%;[Q"=?h:MvrBnnQaJU'Zr~E"DRXZGBg{alKvEG.{njۖc'aSז7(/z9dsa6kpc2j{\_UKuIOHnJBfeQ3LF Ο-q.t'GNJL@KǬNPاKȝѴ#tH7:2-,7ݻ??~g9ig#kO^^C%֝UIooX+}ISoqwk݁6Ŭlˋ[6j9xp$~Zn%i+mދ; [` V<i]5SX'ls+9a$F1B g찆浅vVo__ͯxkUݺ{NWF{eЎˍ ${#= }ڳ꓀X]~}9;/=n^zܼy.q6V.T$BG^ .1*dKFq52K+,5F丆|s^-5""S`~0Z^,&PDj#=NH~MOj"DDɽ%誊Q*e 9_:֋(R9N!FJ8%". @8F YD|>Ee !Up&+I읍Z)ӈvtԂ}&ޚ-cgB8Ф3pҦ^q9Ije%~V(W#d0%^ &"ːV4`>PʙꅣJVz7d=Ԉi*"G$C=,M&+C))fk $-UPiY*~^rpnϪܶ hW7"W,BAN*~X{, !Λ⶚EQm@j\;涎#֩9sw\vUkG~YTs!e))("C'I;$xp.[9&0+(xaI*&K11u/r>0&nkk^XGZrԆ"[〩Fascښ39zLqe.\)Iֈw>%!q!2D10zTZ+-)5hm{/0c.tJ0lW_G9^1'rJ£ mFw4&!|(I`"+Ѡp ZxVhΖ% ^NZu'˜ҨU_H:k\pl$fxЙ";>/3 _#>O.PtQ8Fz+rͅǫfcļ(%LDWx9e-Iww]E.墙ph̛1;;`\1&%^AɜTYDbqI3 O{VI`N%H:Dy-9[!Fyw+d IpұFiuH j5ns-C,W0phr}U,Z{$[56K A H%gzĜwJP(< ?柑{0WF$[%c':1bfrSz̨wt#?t4.8uR$h0L>ɗmCjg0JݷØUEg!WN*)*$Z9 ҢҨׂulEGg]lSԆW+ F"b!hR"sF@G d(Z|VU#Ձ 7xNLPBNܩJusMi[=nڻK1u%JnH0yWX28%M4*$QI?擶QEڴ$ Bб0\:̜]Q(yݶ^EcJ0hN/A5ETHNBh`0{xs_>:@Kˆ1;u5\)XArL_q!G߁s:%|nY؞wVE!Y'qOȒQV:E8ϥ[-K|)(YU_dmZgGF 27ѕ bV¹.fb\ֈW⟌1-QaFrA[$3,шYgB(b )tT3t`i zFcN516 s.^pN@mѡ&h2e$,xXӴCɑ'һ3<ϭ5W]Y!-MZQ1^G4`mSIݬe|VSW1/(a¡4]<'ͳNj"QR#%hԄ%5r)FMʑPb@MU}9k=<өE=F }vP#J >ѡ=/nGB`B@GUELF`> `@gYB`B@, 0!&W[ើEk&LvʔF͹QR#Y5Ii8%3`FV4jrœFMYӯ?H\rŀFq5jFQR#%hԜ%5r)D O3N[V?(t7P>\d`/J^IG[~0'mi;V ^]NI"V)>r`kaEq5h`;5B {蛊5B `W+z(!h4)/yF 6zUwj <5B``;05B Z!0EM = 8x׽hU?HԤ%52QSrbFMF͌ȒD %(U6YӨqR#kyfdO(+4jڍ@fǩu 57RF-eIikө9ŘWL~O?M~gTg׫Y:s/V?'Y'MgrY޽ <xe Ž=0 "zxvob>[o>OWg٫9^]v~Yܰy|&o^mv+ybgg]庭 wr6?Ot83 d9WkozY/}Wn5e\]Eٻ yj\/^74 S8Y欟VkÉXb"DG,ր8sF]"5z Sy69,O}>i9ý j<-kKV5}9.g~ք;ZFm-WF+msiHX39s6eY5׼ ^h;}hᒷMҰzm/pjfH>G rsgVHtzK_vur__>,0\tj#pHW\7 sdd{NPpwW'g^dջ-O?'3Yş_ᑤRUmo/n>׹i^om/naW+ސh݊#@;wS{zGVlnm#^zֲjlG{6v -Qkѧz裿 -jxf<~0řߓX[dTmN\vϠϷ6ֽÎbfS=3Iw_߮ _qz=,=ju/~J!:0Ϫ/w>i]W_Du'z1N,,s @s̃t?r>B#<(ig ~BP (=n%QFpЙWat?a:"bQp|1of$|bIΛPu(픭scٱbΔr|IO-)wml:M['Ctu:1;hb&m"$vmId3e.= H<`e&<½E-u\k"Neoi4S3Q~lOknܲ0] ~^Ջr~ׄc%m!xv?2Z%lZ*سW_ܵꬮ:,TA:1/jZ7)Һb͋e?5c 꿪u8:OjB:27/go~r0|@}]Oyx'tۛ;{꫿PXڗ[бBA6 y9 1c)=XƔjHzW\n̯(ɦ_z7-m@Fetxw4κN2shF_wka!&zϦM hҖk?9}w}|q~sUVNKz͋~qz~N/}Zs^%k;dᡇkgYq'ϻm>#}a]p|_٬,ZͧquSpJ1 l]QF{սm 4^^3.=&cT;Ze vb-.-@Bwۋ|Ǘch8nj#k"5Jg@ct16KHmY mA})!|5ncK04h#<8"t_)GxCh% y׊Fhϻۺ:?nwHy#UᾘW]m8Z>Ea*RtZ. S[y>}.lH7.6zZvo1}6A]]sq+,ŗwjb/T\K֥YTHj\%)ܙQڭ]~\t7O ?~Bc];!=IyN۲1~XV_)+RV\8;0O5SQoM9AHS55!!յY*K}Te{zf]xԏVp1U eL&]Jي,9'S}dOWzDgԜHwQ_u֝Zs M*R8uڍnvb;`r֮w/fA>h?(ދX%g3[uL(mfD"e4x0/;>?/qQ<9NUN,핗o' ɝj3g1YxGr":AW k:3 MZ'tJgR\yD vl e+u1b*9ۢbP>Ȥ֭M ǀL1Ugo$E"H^l7j[Q' B۶uwnw̐62Oͧ^~VտBӛTi=Ho/so>8Nm`mߒ,2?!]jgAIaMީ)+K*2G+Kmd!U e\c)h m}ͱάlno6{C\qV@_l٥K;vyuߝ}&O3+~Xϼc&7^2y*pLMwm>g잕ʃ.{ jMJJxvўž~qv|71O!>9j@ -Gҋ?=؜ ̧NS]-g3*osO˗Ϸod>[sT%mMK^o薽{hy鏟^}.&#pG NtAy}|U;A{MI>L{C}I}2uztMK[ciR~h+뇶~h+ӕ3\}|Ğ_]8;1'GKQ{'h'˥]}\K*Z qo/wv=_{DUǼU!`R%LuYQڡI<91qHԵ+]zOׇ7[^t܇N@tC|h=yNBgLϨ!gHf9#_H,OOcHkvw}2'"n)/ewҿ)m):_c]bZӫޯň:%hg#:QmՓ k]?=Yu㉈Vx.WG嚧dyjܘ|ELr%,g˕pIǪu'g5dpzY{Ȫg4,ƒ¢j 4[~+v:^/K/rW ΃1/[7zM2=Ye&>b5Z.EVd5@&V2@)mcV;}: Xm= oY0(QtЎ +3\YJ_Mj)؝#׋}Ԃ d#sCv_uӻm?<6*^5,lfh*BgI")[TWHQ/zLwyi85XO\j.DeUu2!.8 Aʛp%ZH %W? dF[aKP%`lT&D.!.>!$)d&] u59񌁄6,Z*Y/^ƅBlR&{eUFurNHPRplI)31#D'D b.6njKrBi dRխsR#|u*%46nG9=hɤY s̹R-mվLn (T8N|v0ӓU,ŀzD_qrE$1$qb,R ȐVŖ8bTI@f*mJp5XkŎvud\H$ OQSz%X$^hM`tOceژ u'ݒsQ1D*B ]!8/(ʯGg%MxmK"AUؠf̦T9*]\ g5Y , OiTwPtDk#U)s<$*;(M)Z#5c+@:ے UU$2ƕ*"erDLk D78!w é@;)g>8Qn.GXRQ?(XVY!P|BnYeTw~]0Zz^徭9i%2O;hJ3üZ D윮*jC! um#ګ]s8Xt,[oprA_Kv;6綡qbD^P5I`A ]T*Òǿ*X$ ̀a#?dWFsm! ofJ"uN٘*c#,8pIypcP-`ś괘|ڦJXrH[ݑSH $5" m )҈ 8bw¦-)J++A"ohY0`?363&IB$*!RZpG8*6`Ӏl9 j P9b )F'{zxj1M^!kYqR ? !='\T[dw1Aw0 ^FtQ ZJb_S 0],OY'X lDMoPZkf:oձԎ; k~C9@FP#'"z?|s]R'W .qŔLRs2"S˹dBWHC0m|~@jg\[wrVR b 6J q$=088HV鹳BhJ #N0f^i0A   .L!șh@jy-a#ADDJl 5jHŗ15XS EՃIR{8j*DrT"BK:aےwLŸ @mp>e k=w=HklU2H$ꃖрvdH 1ȂA=`$20崚T1G ؿ:vIW ?k{L973?a_xIB6~W]k\PqEԾnr_nj/rI>]}z/>\ݶRO6Q⏿+-NoU9ϻd z*}kq̋rCwٵ=|8{ǟPܹ2]tȧƿ3ڧf,Fm:nۭR`LmX^t[C^\St}pQ;vmLz6h/-!E|ʙݯ?A7%O m=3)ݮGSO۰趄r)vI{nښ]d6qwLAvGۘnÊ]tħwܟ]pQ;vUmL[ ݆E%<䕻O1Zdʜ'fIֆ6j}tۮ czn[݆E%<䕻O)}StsJFm:nUe1u۰趄r)?芕O͊~Av;&UlVݮ+BFYC^fx!6qwLA Oy=M:7 +ڋnKx+w!?9ݐbWa]Rgi0]°{ x+w!rz' ]*ڠkQcݶ;51U+ڽr)û?ݘm=3) cWCaE{m y.:SƂ'O{mn6qwLA oyE%<䕻O))Ywm1mx[w¯Fn x+w!3on7[jMG͗Zii;8K3+"͗ڰ{pKR;ftRKIғoz(m!ߗ"ЧW9~L">I]iRG#֋kh~yQ^Ӷ[: bB>t,Ϝҕo+%7g6=C#LV ):Wp(ޕ5qcK;C*?ZTfWg^Rak7ű'>QTL'm>,8;|{~6Z3#vVWf>"hQ /pV]䲨F?+wj/ |XH.c W Crޫo߀iѻU,ے` !!m`5H'MBKrLZ),M)Q!%` 椣'3+?u܌}%>-Rmђod%ONG|{I3F(o2l!osdEqIN0Oa %y}'BfDqv| [ yG4`5>1*~1JEyJVZqrs*9}Z2),eRYw%L哃 ;cs8NҼ++Z9ш;֧u_a5iʩ4m%@̩42|tfE#]9ɰ^ɰ;sf!y3Mb#̰g!33$Ejsd(¨3y0`fBdg BTAkQD[Aɠ(bCyvs$ q~6R$HpF$i}^Zc^Hźh.39 F]Jk5A nJy4O1+9Hi\SwhB;`t0֕x>1ͪ,D֕VβSkCi,;Xw)Vi<DۍSXLJz-I:-Y! V*/BugV3\6W3\$oڷI;T|.%DWhɑ"IC{N񅳖pSq\9,2XHdV 1, ŲC9<n>1}UxHъjGR&;`G('msXbXS`Z#ނ`uJ#Wq߁l8D6z͒4w:bO9ۂ-*b*<-SY>a?-$Y9=ExP1űlYF'吭2CZ8͝Esd9y&8gݘ.9 'ea2:XR,Od7b5mYe@3Ns/099!>@Yt7T@YhFu2錵]%%vfI5{!HGtX6T7XLH;;PϺPy8#R@D]179ܬCNg`&/j:ʡSh,GV_|T|wΧcȷedRFYik& *{c uX,l;ducIi| C _u̜~]HpY\yL+ zL`q-Uм|H'NKp9;_wѕxO~=yrZcj'_-{ w뵈tsv<}7.F]O6h:|Waf{q%ZlyfX[= AK@'\aӏ7*EYIV<:R6Wc ^WW$hNa./磸\Wedh#[ؐLsp8{o~ջz7Lb%-J x ^Z|3FY/Q 7avvݥT3* I-\kMpjh#| ING0;~S\˿#wSGf|(Hof/&la IV APAB5<">دVtWolO1_M[؃x*"QR~74ҷJ=֌ątN=4d>~\nt.3(_h\40JyPX&eeNfXpIY+އDa‡=ف`4= DuiΫZ8{'ŏ1fGfςM)CnZF# S3%Z=gA$|P,#jZU-? quO SDNXdwcF}Q*;YG׈KBE` ٓ'4&TrL&N1%QgF@"럢a =S &O{)J k0y';ҏ0y7b y $7k.oKƧ')^x|Ms"iٯ_ozCUSZxgY-{OO9MeV'/W&)P3! SH6b mLaAN)Dem`kb:B Uxkk۾S9G*&^\Y5(MWA Dq@N“8=xC`beQb $c2pďeW|3 :#ZCdhR}U-|7 l{g~O]k$Hy)9ލ1F%/T#,yxnsU禍">{'NacXIIɍNyP^>ـ*O߬X~]Ok3,ԣ=E+^AydA WoABֽ V l>B74gcOr';[c}:y#7m7}̀.6k2:t v+k Ezh2B *_`%:q7Cz01\C|1?r!YP:kE#3+b zρ%\|'͐C28.!w ӁFJcFk&'!dxyH:UXR`2VGyޔή QYk4}Y/vZV0}/oղ|go>$/c7~YA')!* h!z!L[BG)aPzx<QW pZcesQPT\`ccMA+=?RBJv8u9DjQT`)#+MA0hYR +WŸYE8k\t4:ENʃ"lt-3rr4HQR#tl#W"( CK G5%A+k L] \F2$ #yg .n>gj1K*<8ki4APЀJ% 3Caqq8zV3r8R85h*Jw%Fzc Bo)rBߌco8FA$ry- ̀[R5L%/B 8ia|qC  <30\]PN*3f1rR[DG8(H Wk `r3Id969" ZX/NbUwcYļ0 03Qb(d*0 <}Ps!  n5rVBRO L"Pm \bpM5 ̀SJ-p0R^/#Eɱ6n@F*._Z̀ ~@(R< p?9!N G 5LA F $/@T6bY|`PDxD$Z:l&{%i¨:-Y^;>^!OެtlhxsKs$eN ?pVxjս@TB<$֯:Jk1E41 P7HxϬ0CB AT۳; "6|# @Ţ_V/9Ln]9T5េRg{֭me8(J@i+8ne[Xb}B6\'Nm0%d[i?~c &K?sH>SU]GOqEF+\5 w5WA=6av v%F "A>?<!0%o!9f_n_v1kNd`JHdUW.VkM`8nX5G$G{cw?ߪw4Rٔpe`RD46[0i<1Pţ(Қ)zYQDve$NE  4S 4O<Uaro{+S!9ZL XqBٖe[quZ?--vx$;|J$#=/ocZd gi]lYer}j z|cdq'= p*1._~|{q9_\O_>wE3\AIN3h&Q yZCM^ޢJ2^Ve oq"ԇ(IB(y5HmLéէQ2߳QbէV1Hէ*R& {BQ{PD ^"/ّĔyw>5#־TKJS;w F)W[ +ޓ>dUjm2F}.VG^I\%{N55JAIYi(UlZjmS8.Vg͆Sj7FOUOa3RgTM Hh&m/43m $T Oc(jlAJրc4Z.̒AչӠE#)ċҷtzASJ4|TjM'9>R74P) k^6T#Ӡڈ]g7JOe7żHvD2 *hH Br•*u$$,^s2e8]0w֏춼c<]no;OVqlf>2^,&+{VZT.~yw|{]q{syg)oPih^7l?h_s/Bjêsf3-@J'Gx+-UEpxy욦2!( -0#pOh753CPS炅.sS Q{~9bPg;Ϙ crikߵ1NTKLӉpK1zh;uQ43OQӢHBs7{T~F{"M?Cds3 'n>4a}qye¯-f!ҸYX2mц[wپZ/gef򩙷Cn)uoOB-\^s5Xz;L&B[/l(23-l@JiJshNc>Ww6 $a 3G8򳄦=H0͏S N1sdːPʅ3yJ#D{vrǘru^x0QyJmJN Ԋ;G(u`ea-JL%/S:.9)cLʎ]r ౗壃5Ba*Pg`p(yPq E娊J!އHb+H+xTîݏ{ҢQQq{w;r9*J Fu D5EOKچ4M ]-9AG!g̳lbte:(5Sͳ$2Qc:Ot9?+0tI5uZ5;4)2u{ƱR¾JCb0*qkerJعt "g%2eT6@q#x.XtrG&w7T5Tu9065;>_&t-? _~c~ X\bRkI$v$n1TL !59fif['m>j`=D y"ݎcp3~p3\K|!-MS*0qa0n}<)'yrvWf_WQƋ'Y-8IO/. m~ZyIctk@}aR,+Jg=g$وdCk3z| 8B;Ae R~I)n_ =*ȫ#Bz %SV9!/2Q*d81-Mn<\+K¢+P m?;"+*ѩ5XJqőJgk`j:YEIJf"s ~S 8p` (c4.-=7 `.'H@%]+k#WLFk]{1]OwRE2-F8pFͽIQ,6u$F}Z] :k#DQ@! ׎ ﳁk{ ~֡Q ;M+NP $i2_*[Y6%Jq*JvPI( wXQÄҢZ0GdJ]/2x4c%`jP<-~5v ",{2ҊA{ ˎ|zek4kG1kR (Yo^g:u4F1U?/3AԀM6".7Nwusm K0Tz3J_7?-o߭Bjݎ4L:~2=XI\cN-c0qc jeahdhЈ@Jrd{"sSNI_,Bc{kv/i^KйK!޳jTFrOs-±T$/+mִsHTPˎobpNDMBs9Og41:zz יQfTft(pAKJnD{L% ڍ>ꙵn]Xg:oTIfJ1XH3Ǎ#+$~UwE o~t-ywu&W!bKx:Dm$x=L fg!De%5*ld#^9A§&nlgsHiUʢv*eAq*B^$AYIp8(RR#h,M32Vʪ%ǿƏyue~~-#~5= +w}kٶ'#WfنwljҔrx95hwASB LVqo,j$$-Ӡ=s6iߟrP7?dh -+cKZy+hEgTw;2o|e{z P}OI$xtIfmM6M[Xދ= 59|9f .|bˮKbW@bV{Gš~I֙벝3m7 g)J] h ȴʹ]s^ ВÝPU)~gLۍ$e22{uxy7XwKDtdwQը} A)(d2eW\((4A{>K5= FhPj2v6EuqNWi3DhQ myd[v-XQ֍T+ar4Suo TMwWS!3b˜"q,'z_}==ܞzNwX=Mڳ0z)q`H.4ݼ4.m^\ɘuiD!#It;(Yef4&OZһ5tT t3g9hp`K gm geb,*sTPu-Ϊgv}w˫XҚL`֦X ]X#Ҩ)EOIKAj3 xTh_ 4Fz tnj:>H=OLԯPO#O 4__}ﺳz?m)٪#VڧLzbhwf'hl^,Y`Hgr>NWFTwE?=hq҆/H˅i|;8ť´#$j0KqqzSȑz4&X awܵi,of-KGShV>U|Pq8HP/ҹ^~sW{p^ <LUE-TWL}nU'^+ŠSnVt_M߁vZܟ5ia@'E܍K6$__s^bu~xne M&/&Gwm~A? ޸5Oӧ|],Ӂ}0s>[7;uNb>:=lTt?ہ(7I4䕫hN9u'Cq'n90\-佪usiAKM8#XS)KvSWRuqݧ:4Pjet"`+p ԃ+JJrZX Jtmhvlػ良rbLJTNn0hmDktG- .x8?NZX`}1EWBaXGX1(q:Z[/Qk⊜Z,q1@݇c@h*0}BF`75 Ua- y~}3d3mCgr' -B5$TwSM/8FZ 'pʷsY?1NQ3(pIH(p#g#MCC7?gQt ј@(a!C鉑,sG ׅ ЅX80 [wnl_JgDө2:qHu i"^w :+8Ȇ`BڳPme'A G7̜ЦZ94rl~ff y*ڥS~`ݔ4g˃vB*%@/>knsh+W.DcA~h s/ZTc /=V Ƕnڍm y*:] u~ "  Mc[uP.l "p,ޡ'U?~FKx,- %Iͥi??MMq}:x|JJI)OC)[@ا!NIaHa(ES/s(7rýf;?nSǚpSk!3`]{&VbomorlZ!en[|75zVon8ܝϟoow߮Yiُo'?g)5U|7w}j#'[w?=Ȼ ;|<#wI_PqHX;gByM|*p!1@aQFɳ 4Rq)(S P2Ѣ%BV&fv;Pm\ 6ؠ۴=# nR- Onka_-]ڷW_>>_wcHnכdHqDY9Iye+V,RA>ix8^4箤ffBK/Vz߽f;zR PpGTSnZB5+*,M| D_^x {D2h=H^-AjAL_>]-0BOlpF@X[*oRiϛ& I5?˫߯79O6WvIݦYcDo!t[B*gR|EB/wYӡ7˨˞VFdU 8qu_pg[ͷOn/w4wiwV ,ȅ/J5MF,AL/$cIy!h[-HX  mzZ'[ 0ʫ,0b)R(z(W$UIYDࢬA^Rsjc)I3C?N]/qgRBvI#tbA*`*w, އ`*m ޭHvn䣾-ȫv&Fl烯 ė+'Tb$ÒȈKt45G%2YC: :RL^ּcUJ6"[LXشJ[A-S }(l;Y۴J,%uFO2@N#Ad=r[X /6gjy|ʳ >DBd2!`,&gO)*AZ5j)4*7mSÊ:;)!JNuX &'š;1r)TMk:C> 'UkLNf xT%굨:J1{= pֱ63g}SĄ#`]+|c)k sd P`lsb_[Aӓ5DȆ#lr IOYy2B!B@b_LadE:1RTRcK0F- c>LEAQ.ӊ@Cnp!4x ;)8P4htvOc)>)Z>^e0Ӫr&͢p>K>pw1-VvĖiK!4t6 >f@Z 7м[7d}8JVG[tIEF+%z qLzŰounЛ{u]tiR-^/7>Dm#椳H>\I]ʨV$ZcDM-o8P)YNj{cTۥZABx i1&}E qfSإSM8K!_H{|DbdhE] rPUiJ݈dvv)D0oW*OxW"B.')411ּUh%7[-% iv4h3NehI6F*́!7|j;ـ4~zث7g՛}ͫ5N>䪣6k٩i]&kҖ<$O~eTk#zM\dS-2Ih(mZIǒZזPiX^B&Qb k"pBpg@{$|oz\VKܭilb K @e Qp)ƂA4:ITb&M̡#cݘHU ޳5a|TQTL#)<:i)O{r;{i串 i v`YHRSԓ+B"ЭO_)v"(-jL]{Eo P[$+y=y#SFTk)Z7:QdiehTtMQVxⷓhIZT cSٸ&S6C3YE2$U6 Q FBrg_cjUHu+eOPMz*?l;X-"iväpX%RҲ]j(&-=h-_4S03 -O7Oʪ玀 gյBcji 1бZ FWA j.Ik({'!@@WQ{~W˽"t$dML#y ZZv1p)Oh)Pު*(+UNj^dȴj;^+f+uH7Ah=ob`f2BDƶ 10f3ِb<Ѥw6Z ݺ3_%f1'ϛ=B$ /:W | /W}oMp_~9$Uy_WSJNI1xk!mmFG0!MOBit1kBZ|lz(e+T>Є@G/Xxd:r5Cњ]s[5CђqrHI6GZr ǴA\s<,NnVzLG8uET  {aq,$[[ߜeqY^pi!*IKF\^ҡs J-x.LT/Z{v&͈)NRJזּVTGzO@`Rv\[ IH: [kG\PІJ!.Q(:H^6yhpYomaælfc=GhRF/Q}j.pj}ZU&[Lq9 -LK9i!j)Sf0SSxHJ H@+")ܱɘjy&KH{n@Y\ &$斃i yH(jNBշ7y R#0=kkGwMBvur꣭kcQɪ5k2n]e[N[Kܥaߞٝ9d7:wjaM L_U-`ּ NߡZ8؁èoN8 whü^n$t[&ָoK@l,;B-QLqB%LAQ:EhKI4c%%), `ÔNR.8&_*+:I@ P:nz9L."0ZI)JG0 Fp $p [K~ "/'T1$DUb la쒩؃`Z\(I.![*I3,BT^DFڑ猙$~5H7 EJWɥc쏆=N:PFKE#QHd9#O5C:3#>4Rn@y(B=b.b&Q૤B2q(hUqud(+ƫv_Yg,^͗gzryHmjT3{dvڭ\ڻOn?\_3іWFgW?#DHՇ7;.O>>7( "!!$*PͼxAXZI𖃯</yh={<F+.wfボ} I"u={w c~'Y# [i_j߀-VJ\RHP sh8{v" ߣ$` 4b#)"5t&[L;>zDJJUZb?*%kc QW|=uH2eNu~6r- o9.(]b|_CmC2/(+ϜH15møKymvh= F(X#}`!yǛf;̯m5/ nd2[ gb6VOkܷ!j><|_լF'5*[+"R)pQ~ Qj#n AIr7#92-Lͽe`S -&_~ogu^27\?_SPT+{m?.KҊmMQ^Cgzs࢘lGF:aF_K%E~\DT{ vA2Ym9B}{2@!c D:釵`=YBG޲PGd05dvz{/Y}5/s;r7Lz%)? ےfe޼s6?~.|<)Z=vPZ[Rxrw& e_nfooΛ\>8KUCo29?|PUY+TN֮k0ME w|qqw3yCzM){E, 8gL匉1q3&mkHy1d&֒QB%ar:4[þ/e |ceٻ6y9,pvX,< O3}Š~NIǒ3lIq-)<$X$?I/L.1ҁg^ݲSL)i T]3X yQGSCG QT1rjҭ0 !< * %2oV; F(K5͋=2iyo58-3[477ϳ?MMSK4L7[X`.% hy(FDrY))سMA83^#uᎶxGH:1 ߷W'# wQj,zcgpJﵗD9l֤QE/ڊe:k:ZFBaz*f!E<%Z՗Lg+ͳ+p=hSQQTx eބ >BU5or3} N%(k֑nQ Rekθ 4J5ڊ08:^WdH$X%slqD˂ !RεAϲqEܱ2bvzB!Fg+NCkf1"HVY[S)4cnv%fX\m4_;w ұHr6]4%0Fc[l4pt>̕zFjh.ܰ%+b6NSk̦N EK|cub%+p";l1FX2rRsD4~W+8 $OG~pvD޷x6E6yp1:haM & GQV^ِS/F̴vyn**bl~IEw,RyhYi{іk 0]د/8dJ9KK:m#?[a8AlYҨJ&RFUH|[sUspPj+vDc ؘ]ZM)`aiMgǺ?8|V_C<[`8k7 Wf7}yv Bm^;0ӑУݰqn؛kD@ՍZO Ҋ\1T |88y9.v-)^%lO @))BN)'OXgU|Zy7p.mW8.κ) /?\_ו3| VpH.1v:'k>sDەH_đЊ3<Ɂ 8 lx/lq-)Yz `k+1 6^,7' jDDXfɳ+o)>XPE^ڶaZń=Bq_xWe wE]p|сikX~L!<ƾAѢ|fe x@Fa#hH&7a , (RڳjD}88hOvhXDȩ )A>>fMRT!UU]k-n,걶\̐5/D!FZ-Or-q_(B)0Ga,L*afCq)'O{C"ώum;E:ej֔o-iXbο_Z?G;vy.ZԪ( =5../ MEXZWKkɍ?k?]+Niץ"۽1vm#9B@ݣ,JCp~HTSz>}:C;zj@{6ZgZ#[,fԦK ߠ޲̘F$O,ec=[/>|‚Ԙ?PGZ{Cf lQ|ˋrEIOND0Hp= !Ef3jK)ؾNtDK(.G}Diȉt6< >0IT+p}g2&+hF`˽M[&OS#f'~jo@!Ã<屯 Lq|aRۛ/pKIٓ$^8f/oKQ=%4F=oE K!K}o'&ݱ^\xWh_K']x'`xaۈ@i@89s~?1=xu=WgK ͧ{h}Fb 띹ii9ڽWϥR37':kr;[LOqiq+w[}?_xynח _.<#u޽}tY~oyI7_ɓ#|e|upk MX2S 1EsIiH씂 +]ˊn3ިG[446}cniYCZ\|RҨTt͵rPVJɋL1عf" d¯íR0dmI'GD=u m LkoJ 3S'fp%# FKtsųDix/넬"+՛F~qJ+Gmxlfdm 5v>EG%-O, j42N~= tt٤XCJfZEߺo.v'Qb | 311_SR *%"ɜ7ΥIE#eNQ5%r0E51_<*P B՞')E7IBC֝BGwn *RENޠX4cv1 n뾸m]/aP'M7@M)8-RhqP- Ovj1B%oA27xDf8RjNB6S2N/lٞEվĘ+#Ӵg?} !RH BX]5| ź|}2*ݮpH_g%m  86- !iֺ@$Ւ<.o;1tvѾd&xR+ R7?^ DkqƲ/X@m.!{v6AOUEݘ~h$u }/Kj"6*s.//_5zn+UB?PPN9OXp*<4*Z KR IqPs Rp\X⦔J1,+B}Z^S=DUDZG- P gN KY) EIgiٞ_RPW*N뺹Pc<'ҨTZRNqX 'fZ36 \9t!Ui0+('I?~ʊ39o}"֨p(kp9zc-+nGHpЕD>2V%u98缪tjkdm+i1+0zRcqfBqJÕ6CH08/u˼f4 NbtO42~Wْw)lI%5""D#ԝy #&}w;i酜<إldV2dOz=%u{ 5JAB}9ck5~}amwj-ef{C."%s?>˽v7b{'iΔ6-)z,Gzzbt-VM`tTpӣ]LeĎc.,]@lqfαQόNW>c20bI\HdzfQft֣cxٛSVAw{ch0S@+I"~ A~"J H߀ H߲iem,`(SAÙJ42@ЖcY ɡE N]TN*L1p%!iD-9tICM'jNjF%jRjbx=APKP,QKQ3gZrdQ{dA+'cdD#褒sQE255R  )1.@ ¡T;Å&ˢkGD[5W\kԺ} lz됺ɉ2ZbڴQAvmz6ym{|_k>(z}Y;Mzuws`WnjYܭoŷsKܺu?Wwv =f_m(h~ZSo[7bEN>Pg =~7D>oӿFFNGf M"ֳ5%?MmLhMQ[{1aBA餖&a֭[zV@Brݳ)hSh"o{0IoN_[nOPv^ґ9/o^^_owEk]7*űl]TT ZC%u8VB ^hK Qʗ6Ӣ Є_@ R3>K Yk3)-&ĸJT |JRPBJXEy(+IXO,,vd\vdM@wEҙ^Ia8]9v9v:.gJ %'/ ( frA&FLIhh\"xB GO#Q_2Eix)׼Ww]]qy60?,mIu5qS,߆l!jI=~ +j5~#g07)^bzm5F&x,B%#erO/YsdW7ʭWپ76R[=FzN-, =|T /ɷ="qZ-EP#ƹpuX SGP'aKp&xZ`uӣ]Lz1`fEw>|>:kO'xݰq1qǢs(Fds (Y{^#Ӧ"D%kC CF" E:C/ @1ʆĂ.:Ţ֠PȮ0<Ӡ05pB` hs&TC "8(@Wb)Y3RRP(Jǥ)5 \ #5iV}:wquojjךjk/p;x/=ǡ]qjY}^4{L=g,%C=7ܧq$vD{os66Dx@bJ#eZ VkA. AM&^X/'4ULiRiZңmڧ12Y2P(HV 6(6eUp[h^Z߯`ʤbAZFLʈXjh(P&%|&dSB~>n,["AXFw.BN[rd{dYPBrM#Fʹ2ZLR112b۔h'ڻ%74Ի尐D۔&\waK/I܅ĝ`gaɎ#ZWXO~?)-3E-kBk&+< *eȡHȷZ@?8 k+wP=>G\M+5'@.k_2f 9%lV A0w(\Zc%>19"L4ބ{ZA8 >4Cb  )UãˊD`sq<\Vd<ֲˊİŀ=Nӣ]LW_bI[9t}# j{- &]!_kζ5Of]?ެp5!/^_( ~ɱp {ĆK F^:l |ܬfU~1e}_6>~fm9Нc&5_rGά q! "9e9todݿL*Aܿec ]҅a-*] /&?bk?fqRF˗Î6#{t7_ 9TUu6NJ:RƗG:Dױk_ O҈xq4=e/ ҜJ*Ĩ$诹d|ݑpE@ EkX%C1dd @,Ia'oW%utʑ.$9y^6'[l\6,mLfV^=W< r 8A)UxFQ M`a ^#۞d3:Ŏf׮<6. WFz}[Clۙ=wT>-]A@* ረVNZ-Xb %{/IThZtPc`ZJ2YRZ9NHJFZ M߾_揅Vi=D?Cv4=}Z-~m߲Q߳>z`ݼk>߫,Y3b|IE}Zovx}U ~J Bis]7ى~nO^Wn R M>R9(1~Ih|ְ(+Iͩ5X8tI$hR[EIYt-ӲWU1 `qJZ8eY)WF,~ Xi`JA*6_ZcT|p 熂:ZQc2Ths%?E!еVm8hAѲ!?0)cΐct1w?—޾rր ȌT,q RzXAW#0cƓG0waN"cM4DΌ%gm!J?J}zt)+A`GV 2[tpN/5M2\T%nJPZd<]EzyԠMtt3hO1-_lrt^}8IÑ~cƓOY  gYyx=(v EP%ưn$4]W5V_bjWآR"IJʬAJNTk/VuA(Ԡ ~Bkk];RIJ@eMj* ueEE lDUq-MJ9mYՌC-V. r߂B=?ym%jnІRlWqo=bM ֩Gl+0Q앣,t*ZVkQ!&WzbuFs^߈Zlr41v9-T> v1%PB،VziVz/8  %O2"p>!)Vpa>+M&\(䓺6ڢu0M[;Pčɽ_B.Mkl=I3ޓ(9#dfM1l !odrԜMݞecl'v)S{眻”D6Ɛq.TiU:/cUi+-)8QG- E4T$ƙBRVߦd?{q /AC i9'/1}]+J.kA&uR3ww5>R[fP<`8u)"h;p4)<2%0^"s*Z'TVxN3bw\8xJj> N )3 ƙS)HouK鿜sΐSkڮɣDɁè8@24hF'C搄̾tG:DeZ$ȝ 0 됐&I풃M;mgHibh)f"Ic%iK.C 3Sy͝RA089qr_XZi\ŧN<`XB4q"W4ɉ3F qEb˔- &Rbs(FJJEA& @{AIJJfd#"%DKم񠸥D6be\RSU ^%YVL@PRd(!ONzpalpfnpO¨Cێ]jmRUq%e?}VNk#Z=c{ȾM@m [H6%zE8t0$~>? ńP 2VuҚ1{ E4K{Y))B('t2IJ ώWi]PI 8Vl5T6z1\wWcW5=NXVƠBj%z1٘R\AKAnAy8͎8zRL @pS,a:d֧ WXDB̀YlVa Ǖ]Pp%O.SYoyTk~wݜ]OiV1zwOaFsVWrhO+2!&i#,DVSm, !䲇p XC8-@IQLP:3KZu2L y̎q<ygȢ'ph Zk!mD墧GtQe Νrc'KؙN5Z\LVHkL[8YwKO玀 V lrv# ,: Y2CZG@>YP-j68bsnYτS4(|d`f 3a& ()#p8IGvdKV'H){2L-D>Q|+Q@K_U"0_?89B{ bSHI͖PH V']fYH"wc`* % /{oG,8Aa&(OS &` '&DK*7%(ц#*$ :eJ8چ(i CgSIH,8{>2 +ِYE̥xʑ)39 ZrJ/Zҟu^-p+hy8UUFF-YkŕkaC6'\:AX&r.i!,Y솾"eL15֔Np JL 9)&np2+WG[2i)Rꃧ$.TR)Bj ӴF0<D8nqo@mZ )om[xmpb#Iovv(8VmqKM鹠_a7/DE۟Q_wfN߽nvAS o%"kz_/Rxے?>"$}?} Ro7C7wk!tkHߺy**Os}GpP dM'0`;ܽ*YiIsEYcT(YG)Z[h6k>I KIGth I7t#~qּt4 l$=p:raKhVCʿJ͖.X^Pǹ $eϏb}~l18p(Dz E։%SBtf˺5HQmx~Xj=w77db2u~S:ߔ#i52?-9)5>O+ /8kzZz-gUz^K |d%RNw͖yX:8Y]&R%KN4֏~rK٣Ą6si,xd`Pi2dAN}ϬcJM{ݎJ; 8Qrs:zv&e#tESϘdKaDh$沨*V mX?$9G/\z1"R2> ֑oI!)%{^j`$rU9|WJJcO1fp_nv]) ę;^2V ՙ.2E3CFd(:s_\mJ,]2T.7Ljy㥏l5ٺ_pŷ(]nίnB5,ws?J`:eX6onmo0ŠB5]3%F[vz`ey7 yB!_=҆cy rMW\C~PzxWӯeF\Qi- k;ѵR״~LY\}.]WFvӘyN>|t=~39vʊwT$u Tckl#hhkt|"E _)#0tlP uA(AAoZ; BӨevj+ K͘6E8p0g[an>])oݞSV8TSNHd,C5J<\Z0B)lG+5 S = NFGe 0 =R#$#Zo^R1i4aa"V,⨜r+iqjCU[JaR"K?òp!rHNfdFrk!'cH/q LI,3zmU_J39,L^Xi >H0|Ħ)aXJȸb{]T&EA)#6\0KVt/_WH)71VҲ\pVMؠ/ zuڳ;w[ ]hŵũD[[Ô1J9*30PiR_4aV$&Cֵ7I\:vSZ6JoS\q6%V{V.zzx<ު62ʪq9Uio8ުZZ1=oU5j.ƾxZhTfi˖Uim߼UmdCBܒi< Jmv)'q~n^OmI>oI 0$Ekpb]:"sv [ru@ yFW6)lS狏ukkްc rX=^ROض-1 ="o yq\-$TS,S664uMb$4#T;P7=54 ZW\ʊ{K]ߗaO_2]ٕ/Κ?V=lxA-BKTƌzl?dJN*빕qkZͽ,l\ETVJ.75vjl} LzzWӁ[ )hL֪TwOTPv|SSRU GɏJM ´.bdjuch4m  ! NQ[H1VGH3o~FN$T)JAԷsi`=ѷ:+XLSmq$xպ2 f:k |P/" Jy6kyTE1kI%0 2*@ʣa دIB;.#qk=JH3m$I\ոyl 9w==POstiR1Πd$rc[&^B?$lL>4GUjDko_JcJ4G{wFUhvNhkR? p\aJ52;*QoH)F8@QP?@>1IB r2G]+KSUKUWJ$a5fR}S_F0QE.z zx4vڂFpp:n=x樍iǣ9W]Q ?{ȍW $Al7ZYj!qjoU3y%SuH.FTIa]Yd<$$Ydbҩ61戣Rrxx^NB&tR"'F)6#uT`Y@pֳ@%6Ճ3.Wg ?I5[Lf#. !?ª7_wǛK=L§Tu/2@_0kEok 3tO`EOa@<ѿ4i4(*K=հ!.8ShR$MdEv¸9uSUIZˢk-AUTVjtrg^&;CSGvp[әlִ(yxKaJ:$ԙCv9V_>Qbƛ: \ShÎ}..XEl U"͵[Pe>uhovyř\Moo-R!#[\&J[\pp 2]zp@˞ ]8vHQ.  Bj!\@b#t!Ajk uZN]_yi$b) 0˿Q{v]z-R??~iՙwvf??{zzBE6wSukҴ^?Y??vmEn] ޙ.(|בT!PAEi Ue4­TdM%d%FPUO>tx/~KCVc?~2ޭ~z6tݦow\>W?h٤it:-iפi*1w ZUX+TYV $+Q**m 5PcIKoƒc[Ke=-@[$OҲ`в0cZw H'QFeQώ^lp8_Ȇd7 }IhYQj?{Xt2g;Ӳ&#onc GXFeS&ewc&e.A"/Ȩ]5->d<>r4#hx6ְPf% mh C~( RШq%q^ |RܚPpzKEM<5xY:զg1 ;[ iYýhZ*/T,y +jҽ Qצ <TX:s'`gکVA1anvQͪzN6?އyӷŏV/@l"$niRHQտ=nt{+|%'!VZv#5=*ufzm=IyC$uwI_y%=?ZFVos>ڽU_4n4ix+x\Bd )&%ݠbxX`dy J(ʦ'q7}f5v!Y-v J#g{h=qd&z L6;=ZCNepk1Xi9q88 4!Z os`m1~0X~8,Ŕ>fP4%r/z8aq:F_UK@I2  H=vO켋*RxQjMնZ6k] vke4.wuXMУH#{GΛ#y gl$ƈuoN8R.w4i\ʞNVCpŨىuqjaqd`r#2ڌJo{`!_ LEL UY1ު_\.7𤘍z'PW }s0rqFE2b]Nrfz~nZx^eFuM>TBL0u/ixuE&g\NV tZ6m1u*A%%4(D%QF@?m>)Zcۢԍ(L؏l.Eg}8"#*!TRӮ?0"FIUQc[RAGmM? ;mKog@ ku'%oƒcKjtozF߃m3| pxF|}(dh/ (03,FxZWlHj=(F_L?2z,:@}:ҌzcԹ9WI}}xF/KP6%5>ƮRe6s4+H x>R&Wl`-KӢ͒%92vPɄAMt/jMxx7>$Hc F_2&gq dBF`'ѷ{:Nj1;(%SmbK'ҒR*PL9 Hrm.]iIbo?1!,  ΘEItf5"(4ˬFfQNmYB aL[`-p%Ss8Vr >ބ_9dzw>0r$W:%e|E #iY]y7LAL,l/e -+z8C /7#!Ѳxv8-S`MO hiőrP/ʵ6Z;4kpJemg6zO >0uZַs-Jxs!Nckȹ85.J%P)JnJ0:cfG98&RF Ȋf ("S?oyva#5D:ʄWw XRJE]n W \u;|XMiA%hieOu0H/L˺pJ̴/5\M45'*L7bW!jaar Ef {EnCh*l4Ua*ik(jt-hS#)j/ƶF(T؏lҴii*]ާ Q'KLQZY65(`MU.al$]֧@%)Zrl|aIiI ;re,ݱM4e6}}2xpoCM1QKƉ~D$r4MGGɨ1 I5%dVg&B=¯E hZ ,J6ܰFfE)T:<ǰɊ9[S= N9b^~;N,(֨t'O~Jה'?1O)U<^O EJyL-LS:զ&?qt~f!?J(Դ!NZ ri+bsesҠOl]j&'?%ݖ8 ՛,ADIHʛ& )nVRN$!1*%\ODz" ۙ6(<Զ7edT)tnT}`᰼̤&fBB4%?k)z^htSU"YI~3Ny\ +AZ)6BF7O#Ur'Ҭ%m$oZ۪]krźpTZEmu%}[F'g$(s(C;+He.wBq9R+JAI'/Q01ϵ:-GN|m :\'<%HѺ:Ap4=)hp\>Թ*jɏ:wة@<]u#QK.z=uMuM@ONt]w T4*_ c@LDNV tZ6mE(nZWUmS*?DV` 56Rxw&hhzA& mHԘФ1)1cWsHq`iĦhZb+KX tM#deMCR ϒt*3 -9E EmH-_1r~,l^_wN_Lmw%DhHtc<>]nGFY /!s2n>~TWBΧ`?}χ+Ee>Ͽ|S%#hc/hgmٜFjI`XQk?yt4Alzfc> k!>g7G5Ggqfqޒu̙/tj";̑4v?9n_25| ,ö'34+ a ѮRfҴDy1Fmhak&w"!daŭP Q'aTKo>xS:&>1rs\ܧU<^OECaqjxjqdsU` ' rXIs^o4pA<-K9Z;|zVXf# @B?ª7_wǛK=L§FUκ}q*Qe #Q]yvSbx6KcHlT=,hlTUrX Jw92&V ,8WBO#]Uf#|xu,6@W(-亰ZZARK5V':41!J PRw1@|?QbƛIg W3%@E4Y b4(G='&ǭҾ)y)Hy|$ZҐ 鲝 MWuJd.ȿKvO3]}uPWӋbә.r~4鲧'ÎN*,˜ _@4-Ȁu]TH2MIqڐ;XaʈJnC+݆V>34![idHR+S(IZ´ԥIJlMSUTCQ̬?h٤i¿TvMz3vŐj]j_Tt*ɯ[Skk+Li*J]UU%V^U8?wKmc, F8\`̬PB2$O@eaƨ 2};~)汲يGzlEvҗMʊSޣӨ, :$&#/ WnCe`sɞ?2*HHH: R%Ee.20>U&Q: L^Q2YZU2@,YܺZCP9=2giW4 r4_# 5",މt,hJ,ZWBMDvO'Wx1Be+p*ڇ7ǣ61*t`J&lJIa g Y)/\_Iya3|mV[5plV[h199+֬E8sݧY` 9nOijoQb#G2oɌyÉ#ʣsu*ҩ3iڠR 90qT0k`᰼F.LHLD0  f`YTVgj{xZB2NIr%i>>F^]F8)`.@\U(uY6nkR)T:.YvT% <@8Q2 ;IvVa(_Z[\gb2;/u_7.:7'ȚR%dy h3Izu6յ)!gSǍW0:e;mAײw']=*)+^;x55!){QΈsU.}jy2%G@ 0=]V" ff`]C! Fa^hl5XtZ6mh~+3w 5P*]%*a[Y8]VTj݂"!juU3~hkD6iؠ$Ť>&%f!7)$:E2X9Z5DJu5ԥT3 Diƺ0-R8D{s9ƒ$ ۞``jI,-`Q@~wH2j,YDE};yӎB|"e2c0XQj?yXt3x>k=p7V:KQ7QD 6)(EvG4#kNrB;8ra`d&$" aȫ $^;Iv=dHX~`='M֟EVGt M lhr!Н,j1а=f;5(oԮmmCCt9ּ}taS'Nb(At,^`Smr K-Ԥa;U<]0XE5Ff629=Tщ85cbI6%Z$\5)eNcN%)P÷TQ4LO>JQ4q4%tf8ˬ0 Qnj +74f3Z?pBϱ:fz}y܅n>݅K(|::aTzwp싼Qh^L$0*޽Mx*G'q,0*z*FP%%z]Q *80Q`MKpq bE,e6 -ɦ)- Ig^&Cv[VR (/YN."Rw;O8Z-Ja|%5;*J(ߕ D12Σ` ^8Vi“Ǽ<(:?a] Jԫ2a]jKaa.;QP']كU C˺\{UV q.rhy^مXPM  wtFiYW_j:4{yӲZmhކVzo}ZhLY #[[iP+lji Gh ))hjmL%jѴEMڡ.)Meפi*1cW i_U(Ғ5LoZ%Mio/"EEUTMT)ltQJ*U`s ”c,%xв=M %TL -- 3eyQs +jDˣeQώ'^j/|m#7eEut=<-e#ZGCZCR5m}GN?2*C`KJ:2l_IiY'Wd6<qJ"t]-|xB49nnhZ'L2TYԜePcxd"ꄴaMOˊ[SQ'eTkQHGMtMN.-k,M=\ŘQT[@(ʤeSmbZG'D2v*.%jRK')IQTt@5fS*zF4mPxLUJ Cyo%ʭJ۫`ck{ 0 Ѡ)~I=E=|`yT\>voFTt1dл08FmzȦia.sʔSٴ-L)@W7dLg^*;CGv`fV҈so]␀UdvueI_$%TrZ䋦XgW吇Ggw?:K8 uԡK=y5>/]|T2\%Y}N+sA+^&h#\Vsc.d=:U0\t^.ʁDv`Rc:QKu"&x,))cWC FLیW Y!S,u UK(JRU (-.knk$i0ߍ%OJP")шۓ*m+**r:\ Vv%Ie0G{۞ɮ4@<ȏ3b:Q+A;"CE9 3㩗}<Cd;qfC9U # ,uP:M>H3s%Ey1V(JJ9:0 `{Y)(Gc,3bcR- 5JmsxjFT,MN&'B hV=#+~gRB:>ǰyO9[SQToK27nSҩ6=刣÷/.r4ܙNbLE9b(ZleRҩ65刣Q9&='dsswEN S%_©uRB;0+ݙ qH.A8᯲g5zxV8 iuF[x'nrI2}ٌBR޵!BCliݗ]8]G?>L§FUw8`_5fhJqۦIߢƼ0~$kv҃ؾHƚy ( CʣNCʣJ]j=*;i&MneGEr3r _?cڴZq趡EƶE,͙oDwG">74|dCr#H9.qT#4,C7)F(!u o .틑h!u$`/IX-WNTTk7B%,.؏ J8!h}H8kc͛ǏM}zxAM$mRʮIT"ƮSMΕj()*.tH[JZ%yK{ZK$ ±ߍ%Ƕ1PZX*a]Xe==eaYAƀVv SGˢOXF6$'?LCˊVޣǢ, 92R!-kBO0cxn+rI*cue/nHA[diӘfNr’aZԙZhap2#Р ͊ دޏLa HMt?DZd!KȰSÒ̇90]Rbۿ㜦!;elxK6 xX6FNbL c)K.2,j"Ø:*=ŚztBMȞ$g8[2&GdhRtoWC m\.װ Y`i3{Q! ktB@FmUa7JQiZ[{tZo??CG\/)/Z#J靕 ^8s.q,Udvqe%Հcr r{cR7/zuUiJ (e;A rv\iZ{%vQ6'(i˓.}6GjieOSFTFIP݆ZL$lp8l}!H?k ;2?.2|yS+ 6JoO4 K#KK%@ BFRӲi*2ȢIPc?45GCE6iVh&V1**F䵊ONKICLJo*H[QEY(jlUTP@ ֕k-[mQE±2iVhɱq%a ޶g0Xb1X݌)k/·cbf`KqzoW{+, cXjQ>R$ӣ"UbVzˍ.lUimMȿxb?bg;F exƈ7;LJ֐uX+ȈQ1Lb$p2t]:99#."q()C.յ۫0~ W!X?wN[P1E{hA RYafh NWkm.YH%]g?Ae8M(|*użY | ^_d}n ǮGCrD)p-$@6N dۢ+ҭź QR ua ˒wcI -ˢ0AM؂m@֢u6T[?QS"@i, ΁]ᴷiS(RX9dr X yE][c@PNb/-+lOֿO4 )wfΣX&lp:D9M=Ȼ R7/'}[aTF ti)u__^݃S4}!v?LJ~y,vU?~Dy?9{˛ysnnA)ϻ{ѷǗ?ҍ a ϟ~ost/n9~uͻ_[=[&-Bo*E"Y]F_woFTTBvBq lYȀ(pQPmٍn@nkmɢ zfTe^>PBڕwpknT\n¼ɬ?- {]-juWK *LqݸvAVZB`Q 4 nꚩPuU ᨰ/ U^Y׾*Jga./wՍ7k F'k?uuݧ‹{ۻ?{ru}ƞMݻ؇LJO{s;OoN_-;ny{U/>\W>!e)>(dQc 󵸶 \R;X[@GXխRsR% xRZI(qA1]aJ\ NszJ7׬?!\웾EhbI)q{o1&Kč54cD [%[L_{V0*Wk]n!+н,]K*G ,]}HBjmsY$GeaJ:Ѱ S,A'ũcōᬿtMRKSUoDE2tML9JE(DI6},J6qN$_5)/B'bGqП.%K::/:Tr`2ɬpf|!Vy)qc?Bxk@IJn}m@i[{ȭ`Wֺ=Hz| {liYQ@k*>Z`ʱe`Y[wn(M ,, jUOG7889ٱ),k)ukc)K>0ҡ²&BG5dfK3ՓkFr}jAX(ʊAiԪjKjѲxkގUȫْ`dMp :1Fwfr.G$ĿuN8ʘ$U9XZђh0mο:R\GNM&o@.GuD Ln:Q|=@v V)1˓νxpiRyGN)fmTʁv @ &:=3)K TǡuJ2ݕwpk-N2/@yߍ5)-#n |tҡJ4%r.Bqlkm`vHiQ8O}5DŵHt*[ڪmQmE 6@F@F!zRJ~L ,U0&T+Sm/].yF 2I!6{GVI!6cMƘCxpkgդa8[3tʙ_&#[-0!G,]mU>a9Uئß 9x'QjYŒ1GSo6PbJ>ClO'WsAlbe'S66-e#J'6fF†4-l(F'A`S{J' ِPbtfs>6Wb;Q.I/>))`į \ TQO񫠑 ZXBlF)jf_xaPZHqtc?f:㲱Nȭ`XHNvq\(ƒ<$۶g`i$JZ@(Iݟ>3RX]L+],zR+i .ZuO8W򞡙#0>zɜԉ;>R;^p 1u"H3E%h5l'Hcw'&f,HaNL0|9.a]R18g Uv9>ɇWqcܹ˄LuüUB9.+8(sCNNTq 3p\.VA' C| 6J2ZPhg\VH Vg+oa67H L|TXķn 7HEސ_tf7L|Klњ修TI֦nF۵ߋfu j7n /]$MOVMk]uWRe٢XQJQT UM[kuC?Z]ԕoު{78Iv2~zm5~Ϋ\GY* kOi*!wbUESԺMmچJVjm!uZ T;@(O:ƓckO^2r>]wyS??/~Lu z?wvakX |ÿ=~+l>{?~(C+~,G]߿w;Q7^m~_)v3?0O?q緿>;։B>~WQsxmFÿяX=l`EI2!C jVҽEiYq.OWzJ˚sp p. {cX- )-ka٤&#j+C-ͨ'c܊eq_%+]]K3BHN8z0p2G?2W/ͩvTD-gqg8,j3""gMqӲO'WqSѲ"u_0Smj SN(&0M©f jktEJpAq TS$%,Q7 EszZ3Z 6|eV k2ɬ8z >+tA#9|02 ,lVX7J*f`DaX^]mz=+jV#YPK3b3`{2,2!!4%(ˢzi"`y {eo͈b8e58,`"9?}f2)|)`UYنx%UU3h<qYW =0v &OaG $Oy} D=-@t@ <-la̽"nQ;^p9vQiJt;p 휣mםp|plZwjvwʊv ({c{%wKS.Mf;Y2Ttb6 |^vY!^kWíAf_1Xdc$# c-juWKZzݯ} MCԥA& .uIFgƐ)tRaeVi1Xc_4 5Vi^k bU)&% 7h4XEMME$E%VBU,J ,ʪ貥/H)QƓckO^2oc: D39@2`yQp!~(q/Uh0X/EMJ;|$fǸ`5I£8O1XSw6p |FǪl%)koc2X;5d5l,lR?V( Ura4cߛp/G:aXr׶ 1Xasy"8x=N8+jcb3Law\UoDE-zH V:&`EʹDx3HR:f+E& le פȐ `b.L_(F-;3/4.I2S -I_& `'`%]ЈtNॎVۻԱۢy]w{3}z~v?}2{}::aTd^$M۟f: ]840*r>J#/ő;^pB3c'$H¿9s%B~ĕKSGd:\(wwy DKCq;$X%ڪpAjg>f 3Z "R5 ZVPOeEmOsAp u. GcHK:F5dd +2~;B9efAYzʆnej#DVc9K7үnӲT;6EiY1*Z`WiYTeBZVx<}:4tfh$-+jcӲ"tRZ7E .%j\RN`JOK"fj!["=-+v zrrVTPzNVTyNzVTuR'tzR $ů'Gz C0> :㤁 ?J yé4K+>ی* 4 ;׌丅^VH@؋4` ᠩ(`y nˢH;Alvӿhm$m~`2) `aଅm$KV~'Cq5麩(x(uhiyq 6u% h'iSAr8LtV:k3fis*/ y  f wB"*zcgIm^b,Z%믒5yQHB.y%tKB>x95' g(]*Y̜*07v-I yY!I*鮼 ,]p`W>d:/RrhMVܮ{gOp&WQ}m7X_q[$V_yZ/׳'M|ssسI&-n>o|]ߚܟ/޿.W~pjO{᪾)>_/WOl&9>Vוgl]w<ɝ'nsSv{[݃vUEH,D(E.@jB:9東mV_?;%vʼK!3-(P ZJ(uR˶s_B:@QWTT˪h,4ʢ5bHJQ&;W8O"OE 2UamIuYZRUej%R@O*A8[Zbn9z|.e2{ym s@n+\'7} U]wԺ3wI^q.0p>k>oFG7VB6xcyLaxy,) Dx&QȜ*`d9~n=ĭS22Xڍ9{Z)9 ]߅P1 us5 Pqk8*l#h= Nt=9F_cKd0(-?p {mNq3F{(Oe$+j#㽢tb+ {%jTWNPf W:&{EfrG͖rm7Ẽ YRsT=Y'MlOlIvZXlAj*Iu6hB>7`uRRE´:tf꼈8!WY FuO87M)Hg&ɳ Y @PT z<#4K -v^`W8.V DtK[%,ͫ.N I{yQ,AXoaAXQ#$r0]%ddo{R`t4e28:FN~Q^*߂;q8 %MĐ71b7s ͓^) /6[ַ3,>M[߸dº?.JWuw}oM7kD}6\r0J+hb~ [%4¨ K[U\?Ӵ7__7:|ReW] Ї_7+7zͫ?o֟n]TUzq= ,|5֍@ؔ-*lLʲuc4P*Q*,Pat74k,g!iYӲ\ɖE g`+ .(p1 Q[?y)d09-G~E"w1eUnVw6WV"=DyZ} _\>6mK] * 9.7w;{z_7n}\}-ύ6_|YSBA`B+;ą!Xo?/pWwkS}cG oqV(\N:lƸmRF{-9lHed+a!'\Fٜ[}rt-ѯnCZ)CZTR'cΰ_1PcQ8,t e|UZht6x;N9$ʋy7h[5G/u CJ G08'tGGfPԜ'xNw1:i"H-. %j\]Nri.bSeFK!od#st5dRM7 O-s:/RD)G hSzW 6>2~yk)<᜷0Ң@`mX\ ^%% kLH'ň7okˣsm`X[ENfK ߶~aaCܪ@mh֡YzB\ B%#zΰ嗥;we@k4 -AA' Ǫ!o86j##o#ղhT:a1!c}:cQ8wnj Nѹkqv_$~2V'#ub"?Pq:1S^yl`/|Xi+~گ_wWڻwRɌ\kiJײtDE,* XҺN:yT,+YiР4~wf< 1h_nRXL_L[.{M9 TS:)ƱR> Uâɐ=cAd:#O~kӇ>Fw>Oc,vpR±Ygd!W2WH"A ;B-zzƈKO'Wc-F(Az627F'%;w1E42:F'T"4-N)Ბ"dgeL}!?&Ny)`̃6=/$t99&9f)Hc)!Uz$ҶPp(񉔐7:Sx8?~T7 /HU: ;;n6 )&ؚP1kb%U\nY DOosLAGGIAZ"rѣ3)RfE9Κn{MB'=k=F Jq]3bKQʹo<J瀿RKR˜:Hb|x1L̷+o dЙcO~׾^Zzkuv Е%cϨ 1PZ5 ڦe(HJ2ֶ+p@m` "@AUJJvV}8UH-dEck˦2JndKV)]ɢ44 KY[j4 -*ʓdū9 ;7w^oS[S"Ƚ&crEX4Kbq^KA|zdW5(ޠV*?+: rTx>{p]07J4+ԘcQlLaA٤xQlVvR<` I@>~Kas) Α5t]2GWKA έ_Ex+)<,k =e7E{Hldijv)K3xV*i$fC] &Iΰ Q9ZM6qak %899<6Wҩ6:5㺧#@"Հ J'+lhyf%NYr446cBFe3RtBHG~q0rNt)"́NoLj+H|||`ZmI#a -X-e{?NqWtALB93Hc7a3ʫz+b>DAR?FӡgLMD5`jn 9hӝ"[І̑DeAPj)QRikC/׋Մgh|d`P) у;'gg)E;u\3C8YzTXiQ`؂dvSư E;-?@

@FR᪝ r})MΫ*i.2}ĥ&E :^ԩ%촢sSN{=V9ξ Ԝ :Z`b^٭=Y+3e|~ƐO8p?#L 敒^ ZAE`Y]w+E|w>,Y7MS)VB!ɪhRx[SpP(-d+B ˊ}Q,+;C6iRo$tMz2p k[bS1$%W 65m-q6t-*[mk 7@%!Z2EX`eml{ ,aYòE%$]+n2 Ѳ:7v9 |-3oG0 ~Mx\]#R ]#ޙYnR|oOIb󃫗?]|"D;ַfrr\Unz[:>սKu>7?B'|Y=[M<ѧ_hFB6ޅFxe4GTRnj蝷K3.iފboĘtv249|30;aD{;456t=ށ aY eiVeBEErV:' M˅99>xQFkMᐷtn7ZȌIy9C޺cAZzp㉴tM !iia\"/dSh l o ׸pr٬Eh8jZ@ @}R(1k{R[G_ F֟7 ho +/=:_Z#J-C1pQU&"`J1!!F9HԐ6}E^Ïy 1w&SvӻCi;OUvo63'<PyQb Pȉ)m@uϑ񈖲K𥾾}l]9|B+?5޸njq×ju=ᳯΓ]׼\/s><,Y'<=|z==~yztͣkܟ\~-nu٥|)n?{JI1՗kɭ/pcNxzuZ޵ק#^ydkHɾ&kNU2cJv))(Ty7Kt(sB-|m*EMKxjQ>⽛nZ zrs5pK_ӽk;MU_ τM/.mrTeU77VE\S ] SQfYaT{QKq7uԻHx0ů77~amьiWvG`?k.Xލ $Qu̾ qt~]O((Y)p#EGpE29ti8 +m,+~`7ʚ7L5qCQe+l[)]3Zڲ*VE53>FsH ,j[WF2(j(T+Ac4*ʪW[X~݈onNZdIQsݲ`[_8;XR eڔFVikնԌ]BF(a,YNX7K7JuΧJhlV2-((kpnD{v?՗P)>{}y{Ōa^׾% ծR[, jzv^蜟7WؐJ1w4y VۃzXty"@;OsJ؃;03rXr9o+RaPaRޭ@Jd&-J4Vx~jYT%lea 9qφ;hZ:`mvUɠ349`=e~diV-s$eǦ9ehj yak *dEZR ktv:&lT~NxSo05{:15fL%N$J_ &R3I65$2m53aYNlChӃ!+xf-ФkIxzj&uU<I~(σg=-,(QdFP>'i;D}ly<їcE5d{U^cXkI ݼuNn،A79_PØMb>o*%z7 nQ#GC7c=$׌{|asC7iݤ:.aq<-2j*kcV1hr)mji[.V5 Sn7f yq=5<_Y:ԃG!I(pA2 eo`"wtTG?؁w|0 犗U׮x W*MtJR|j\}pRwHF}J|+ʷҫ~+|”V\4U+Jk0ntV_Vʪ:Ej_:\¿/߭q(6/Eϸ oA7뷹.X~NMZ<&MS BORXjJaTmklp-EkKgRht6fUo*KjՇdlq/Ӳֶ=@r%DƄ~fupfs(ef(vV(Wé\ EIFNQлtM@pY3 Ӳęf1pasMjnLj)W€l䣙a׎f|¡WF^0t9^t 3+r ao/ҁ!O @غ qL1!!6EAYQOeQ%&-?7Ր9ZB+)M&O{qvZx䐒1`-RzUrVKSbUݠEE< i= Կ gQP`| _#vN8 "_^Rǥ&$qc͜$ЀGRm3SN۵sj2g?YiM%Ÿ̄Tvs꓊R‹;e؟ <'"`3:; kmTߞ5 +<,j09Ċ#ss,o:B`lnXdB`lR FWm FH !4>uEPKqUF-EA$'zԚY)-[jM  d)LتlK׀$iOGҺPK ̬m F"ud^CHg ;rpF^LdW~}̌@W3P90we؞ 3QPNlӎH&Hqpڦ= 3aSy+X>IA;PXʆ"[7fX+ː([ g.H"2]̊0544iHBVڠV`>aqVneMGc<5yyR$c<O6Eҩ6:O1mL*/HLyn&PM:FfڐtܨD-iPQ649ŴILl r}NupNFA2p<-F LbF=JK=qHŢ'OfuV%yUG/aۻ.?ka S9 <`_5}@)-t:.?^޼ 0.кL"3$Qx\/ki7G'QQ1;1>Ӈ4s[VdXe]$U"/d-TlUy5Y0<"=j]n=N dcT}wQ$|cJ-20Ξ>)N<:+ ~R琘vs~2_!YBj迬}ޔc$|YVq[[?L,N@RnӐbj6'}-' /RCA?n jgs9 *0o7&\PPiðD Nwf@1,k=7 *`R@Am57=}b21iCb;((\.פ*.%Ts:Z}v,TQ's Н) 6׍trS\,JjvLE nOX#" P\ W+"^k5kiLb pzh)ked ٣0qb2@zꑥ0hF(Kք3"fcK} ,yFȎ_UlAhe*/9RpJH-R"\ÂVlUaa\vߨ{w"= uf e\zsineG(qN*=Z$C1Dk,$<1: Y\<Idfپ?W?fnѺ׽ m{l^7k0/Ū}۳lo]:w%j7k-iқ-iқ[>γy  yV UK( N%f{"r)4BK~iSKߗww["Hy`&9~vc]spX>ͽmgoC?۾]+?A4mi!ȇ}P4;:Ϳ(χ4͓#3#)d\BeUc!)HSf[UY1b3J'~3ZxRšTܶ{tL}(}ˊ_hzU>"u?l]_w8Px%4K -gk k ;&*,0oG"+ K9g0*,"if]R K*jRrpMf5fAF#tNĜn`h`!0X^b8mDZ5I!D\)[վ+G%1&7LReM̍Π*BۀN<E *x/L )|>,\X2ҫ ئ qesXH׾JVSB,x5$u2N&ew)H :[SLvSp(C]voGG"]H 9;풻qD!t PMv`"<r@Tu:Vn.RWZ4YmZ qi;2hIn: (Oz;I: E sZ4zy9TF+QT(h%dMEjjeR~*I9jGCP0N&9Ր$=UB^ Հq=a*Z J=TtЛ=~Rlڦn}sgsB17.kwuvcդ%Ⲣ66.+@'c\勅a pYEm)%Ტ)6._6 (΄70Y'73:m|LXaRus2wvNE0p,$'y&hZoϖГ ګ t Aw6%Qk̈\dlz3D|glAF nOsv[߄ Ow:/JJTk<_"w-v-Y%gxFL$ )$1D?H[xkUyEiHA.F/U:!۸溍{!1ZXaJD! SMF:n4d ZjL j2h"o>aw57ٖ[0^}t)bN'j`BN=W=+|b.GTh ڂpʔߘjȔPl1ٸj0 7~Zt-k%a33|ovV|;~XvCfF ypQ'h۲9z6;HK>▥U ysSc鲻 QǍ)4W| uq\q*lKڜ,(%5BkE%=c0ڥ^V(57uEt`\׎U)5rb(vfԅVQP6¤*Nn1FtVD )-scBV4j!{^7k/Ū}wfys:/׹+HU;wlOv%i$i$i$J ?Qp]$di3S5'8T6CY3rYidwRB@Vf_.}z}we:zj2MWZh7 >5[U4v|mھ:nvpFJ dKSiM>]>'Uy. %kemmM.ee*R(%,BUaDP%ēn`Nߌ'V!$LoOв $-_ZV۝&Jډ7$Zw~*6:m]Q|f JSd ]#HՈɰ5W8wo\2`]kIǣ/1՞uq|WL2+g|Sm|ƗZJ4,15ތUI 9Y4=M61p(^CFd&\lo-'檱WvgG\.BDG쇹J\4ԥ̖P7avyQA3$\FkiU\Up{d+A:Eɮhcb *J>;_Zuj<7`v:0fQ39Xav3-L(o;B`8|a0^'lg\zȑΞ9 l]m=LIS%*e;+? ~*G&xtÎSlFZFijvi*b:\0W7= Rs[KrF5ݫGtċ8 [1-:8$sCձlUY *Ҟ M|/^.%؋Fc{+mS$+2a d.zuw]EX2:PrTi8jUpмTJQe,k&@iǒ(,؏?k٤yIi*&cTeإMŦuMSեm=T Fup[QM-UD$ԯƒc[Kۿб `s/+W_/̘Չ3ݭ8.ޒ M {{v-k1T88"p޲u Qa\ɞC m)Pas <*؛ DeQ]L ˠJN9%͈j Kz.WOJf]]n&(4gGtрbϛ"i_l xG(u+r{hD{U1z9bdCz9ToumgQ]ƫ|J QבZʇw<ۜz NKn{4gu>u8b&!U2 [ ^v R`gF+$CýMvmd5ZӦ v]dVy2E>eĴO-c;hik«m\ `0F`kq;' %$־tԶ̨wtoJ߄k 27 ?5KY8nUc,I4m/АB uK4/Ӑɗԉ·NhH j#!P3Ȇ BWխEiH"^ye^阆4sL'aks9 }6LSN3%!SEݩ4EyQ"F a; YU6p9ڕʑ<7v&CS#9:qsk tZ@&ueXsҩ7J\:&:1 Ị_ iHT$ЩOc"΀ N9hH:I4kZRN`բ!3 je?ճY-O9'+:׬a$]6 = Iv$@dn?[E^3+ iQ';xPM: P>d #@]y-LnSg$"~*a@<_)JaoH0rl_DاvhpmX㸝(֠MW&0-Gh%ɭϏ^;u/SOn|LZ * nz#䰬u2\ n|z@5Nݢ2ݕjV^H8M*CKEK+oi,"Vћ}i13Sت.-d\[ mP]iTkTV64qE[3DD_%mCS|[*Upm ԆjCkl]`U{ cIsF[tyRWBPb4$̘Ņ"޲ 2x<6gبdVk=dFv!5YmUrԋ|lV :׬@K"5&b%];'2}xR41!{|bpᘏ: *>*k7*ϴ£J*-ʀ$Hx9 4q]yQ;Ix !$$Mo})zhSUb˟&iX$+0fRI mع'?}Vd4ޤcySr75m lFro]YGG 5:(Jٍحh+J`4Qj yezkI`ܨ$w]nk_7^pd-BMЮS mqdP[kж5kE#T:)X'"ppBYt#=zZ&tۑ/|([pS2XƭuD9`ǖH33Xj:l3ݕհ[kr6ϋ;ۡ5?yCdra߻?:8nwκl¸2ԤiOgwk]+BK5* ò!k%~n~}(/|_}Sp2RK xIZu` Cct6gOD n^Q"8-^o)UvFUjs_&؛I|7Hּ\&/N}[! +c"3!xX+( a(&іդ ]HVsi+prdbmO8:he4ֱZ7Nd4IԁxPcQG^LZj/yǛH-#7By5m}:y\n+싓};'wwr0S׮h\ƕTuE`OS5.*Sj㛖JW*z+lFm=cb/ڢ\ :PUY%5‚̪4"%N1v/9s {6<GC}eS,9npoS?[)_nӿ؁)}3qT(;1vg$m3$+o0?A& t"$G='$Ex+2/^ [2DYAumᢩ7mkU5"b  dd'*hvtr~ ~~CG)vO8|9?ZԱdH'J 2*D 5zEq/ n5Pk_9c%-O:%):ى>KfP!|=.-[>ĭ\'kT!X8I6ZAUQin[]:[G}嚝Q] JA[3^()o-FX[ѶC eKPפ* ͅ } EMSAh%ǒ#;K {%wmBB f(҂6Ō6ӑA~>CRܪP0ݜxgiޒז3tJv7?yU[Hއ*˄{\A7[z6pJC>*&R*}}@1J(FkQhڦC/nH_xRBdu\Czi QNx6z$8wՎӷC Ԛh2չIeP$)9\TTdz)ȉc_:7tK5nU ;ʴQX/MʇQ9l9K"{"GS; ِnghWN0ĸ9ՁFprrLE)9Qœө699QxÃh>*/D|<XAf^TZKĠ5NNOZ񠵄BM ZĘNx7)DJ2`p (|nV#8=9&1R٬nFEYm.=0Q4.P$444+9q13I;_ӑK}mfȉ`)ɉ­h~Gwޙ#ZFe7w?| {!73.|3y4)?kwJ^ ?I50&)peꏁ+1=BۦFW庩 +lS覩*[h@֝0WD$Mƒ jJjpT '\붬V* MJֶؔD޹6rcyY)4Ov`3¨k'ݑIOݗU%b4&qwxx,;PԶ# .hjַZIy ;PKtZ(\6L!R %)*㡗*΅@տssHz=G{FE(ܗrW-Sؙ ̜:3q َSGЋ&/vvZk+Mlmw7>؁bn#|g2@ev^cv-X~m0wSѿ|:$1$M+qq&MfQ-S%;֐ñG_ ?bxp tg7uxdn5AɅq yht}M*>L˕lϫԜ4^?<<{`?ϵicӬZ.߽<=Mp=+֟6 2-R W,L:H|#򽏎wJNC᪰mR(=퇽b1BĬ sSe\+RTeL@:Y9qd"L(”'21=y&ݧ+0ݏs@ L twSIR}if!t_GHB~%wdb-ez싃^5d 58R0\mA qVMLMO#R='g Sa4Tv^cAYm /!BP΁HLJT(X"+GɔZje/n6/+)RP0|+]eVZYײȳ27eFՐ3`!^RCt^آN6}0kB7Sò٤(ٻ?o3eժ?M[[ ޼}y?oƿ&$i ԟovhbtzUighZvk\?'7]VfrN|\jomi=-l,冴3P &Ԧ O~Dgr-.M' mj2Z\QmrM4M)|߻)I)n>pi:enzHh.nu݆h!D{)HMu%|nJXlnrK* x"5Ιx"BG%*&MD$t&+'Ŏ#H8[\%ILdg& <9 /I Pkd$OJa<ǵ EJ| T$*A'=%_v8֨^RkQhI{Ue a'[d)/"l=&oz/|Fu"'_} Z1P'_g=!k"L≻Bq&(jɻ]ڦKRiKgyYsUUelus&K%lu4jC MĤEmIGOk=q^"NpF=ÑS1iL.&FMGqv7gM1iG``y;Ur榖Kk(eڝrz>{KZ$ERY`Ay W}O_}'ю#sdw}wj stU_rje#(p{bIst;yIh]=Ձݕw2Wzbp4Q:*cc.åmy8 8e3YY*56*s5hJ ǬdLL_06DnP avN56X~FOD4SnRڶL10ϙsV*e]v"2"^u׆OEͯ=vHO8#C$Z?\y N]C$B8I3)f}-@R8 n G %“hJE,5 PI8QtTZ5 ňTX.r`je+ >*/;Od/|F(CM}(s:-祈F :]y Cxާ{vx.g\НGJvy^~Zc~ &B!8FΦ9IE#|v{F(P(L$5/D>b3)V xrW}Sxj t1kڦlHsm|ojcd,%$JxP|TVtŊ=[4Z7hkѵ hkF)i5II~hSpbi׮F5uU!61WKs%e.sD畎sx;W&)?S$eݕ0~,P)]fuPfs&6%'O@K#ؐL|Iy@M;? eIǤKIOe/ĪVY [=FHg 1Ƕ{<6[{"%=cc*[ؼԉ0G,tqK 0|{%KDEuowM'Cw}}|57wy~p0]lPp'7+X4鍰"q LE8%gcUNT7;vN\igGe,+5v[p&WDdP/I[06@E*'ưeura2?DHgrQ].Dl/=rM "-)Ғ 0ξ]:|]{zfwSmf8a?3}rM"'Cʃ7-OσQAIӨ4#96Gt4(@=,t(K`YNuԬZ Ll008utQǩXQSJ,}/]UΥʪ+umSMP-wY5 wW> +^3Sۗ\81ʶR'ȳ0#׋;gezaJs2W*яB^0+(u|t nV@/m ѩh^nQ  ̌E3 53ؙ%>(2gDˀvb*iZM1fc=! 41dc蔟zq|dXwd֙‰Iv f\OE_$0N`p+XvYD'ٓͺ F m݉KRy NN}Z.->DJB[_CdǾokZWع⠷a"ؚatW>`lfu;+Z_" eN.(aVkVk0Ѓ v-Xnuji_%փXL۠qR9k]r?4*a?h6Iۛ+J- 5}Uؕ7ʷAXG`Үj^T 2j5LikS%?Xe)b:/Pɱ-rր6+~=F,99vEx䕛b3aCy_,ɗZ$϶$'ag{ 7w_=$fIF.upGN=k&V z +_?95ͩ ƮhJ¶ L[/Hk¶ڧ)Q8p`C\B%"n(5C}Xد'X8gXmK?G^ۖ-om[εɱmYn9 mLA=1-Q"$^~t!l\l`aߪLr}cb\OVq/rq^Y75Q|B&(ɧQ5c(&+)o8_ɨFV] ı/؛B;֡ܨU ;:= ᛟcg,_ߦM%=F˃s]`3\iC1vaPP/Q`hP]rL HHF.eV»2! *I?s]^ 4}Waڰ-ݜ=y)eer|\Y`rN(Vn$yR~bh؞05qeI],wHzo&Ifϸ%Ոkiîk^aVA|$0^ ZWʸ~Aĩ :AX:mM.Q/gAuitÀ)Q !ʮ+.AY? N#{&gˡa][WeKa] #s PiX׹@wySyGֲ+Efcz \kU-kojM*tcQ:_9F ME 58סEn|dE4t;X{kTOh}hW{jon-甕c7eɬ*;m9E)u>޾x^a4d~Fb7p:nݦޯsj>)}PkC^0˨!Wk05Go1 2^ɯ=}dә߸|(N>i3:%:`:^X,ur[*M7[#S [|)ǣs31VwaՔi]V@ցkBmjċM[׭b$i1?c(ّTD& KJLjhЕаHǀk].kW_Z{!$B|/ŎUF4$D\y1V[7z?K-ϒ@ӗاO9͐nIiI)SÖC Vb߈Ch|i̇VPW 7t b*LӒ@?[RfoEIҽ؊Ɗ9u%Jx}888rƞ;컰(4d.TC^U缭]AiQ|eL]V\etl]52㦏tg& ^ݱedt;1GejEJ;]yqz|S..+c:bosL@k]>n~]~(ޓ7}l˳ &K[~||/o3۞^pps N '%u[Ǣ, V:*!1y"7]ҩyyNeDE ~'9EfF-J@JB L[)/H+ +QWs"eX"l!YRk=[vs#K\ ,/Ǟ6*TOМ'.G˛smr_[ `9OC2o..DG5\նww^2\-෠Q|>ZFsl_m "/]2p&l{NQH@Edz+dgYE)̳) H/yлc*@}ؗRF+LJ,}3,|3CH˽2Sy&ĥHWq4*͔o|@Q ZJeډ_9,+{<=,2%w% R45n$(xS [zU)3LI_-ɭaJ/0qEӾʻxYы:֤}%ٓs'AW.+͛接7M:?q/y+BS'wr_{~й!k D'PB_YA?lh0Gn.\YWع"\f82agf]'a~bK=(r[n/xua?f%u|}\J&fxm3VMJ67lThMk;uGۇ6y`?څJWuh *U3֮tܸ+cMeJ0vWl N `/\v[$㦱1ڑ{| _nj69z׭m=~LyɒgtL{-YZ;#/pSW$(WrPx9M^[=x~ z$=| SnrbkrM>v{zqOc*MWY/[,#bu'W7-"Ӡ9EJMXt28= јOF/;dI`kmE؍݉M6ҦV#$4%4Q`M-hG. 58tDRQ5ڶU(5AnMժIا"=vf.L; /]*o_;%!8{f[`Hl55RKD'PsqVW'){~:,X,o:683r|eHXn6e AϸM֟"YYkZMJF`H)deԍkM+4FMfoK,ëR[P4 OB Kv[vW{BϙϚǷo?UoU5]T7Ozޒ>߾x5A)3h_~ u׾~h_yoi͘5V]ugBzl6[R'~2u~NnölRFŌBƴ):yptޑ쌣[ BN78ی״#zth6F &զA>݄gbmtbf\xM{qTSn h!WD_ Ү|YxTհ]QUCp^yaH˅{Jxew%&My'*wrJ f/(B~"ޤƑ8ONԁ{p%IaFl"*@և}.2nf䃬My Grx%mCϴ W$$/cDًCM/ʐJҿ3,LfRxzN2SaQ#s(]njYףF'6 ƫM}_@.LD5$r uYf cMR,۵=(T"DJ~]KDQeK` L\L@|gay'"c`Z*|$Hքsqװ~HsY #h@IMUG: ,ӑ KLvAߊRt<U.쉷 Ś2Wf:D4gD$_u6&ʗ|my|rjObErs'Ls`|G";m)Ҥ͗IkИ<VWF_45|MNk+/4RE&Mt[}wH )tfk4eMz] *٘dUciPBN5@ͱoƒ6@֑s-쯪XDkVk KTV+pb,}>ζSyN`9u9D1JpVΠd bώG_6$ZTԭoVDEvL@F\ꨴa]03*ޭ1Q)"鎺)"*).La/bLN Үcqخff%kEehaDKzTi)K@hS@2̀|0]ؚjǁ z89.7҄ҩ6:络L7mNb(jX($gSmdrK'% L0]:&ӱdZPN >&_r:uצqS,!UE(]>r (ɝffԵhu[(@u,yyʫjyW"f!,~,3omNv̌~h2,`5.橭/\&+97KVmVmt?s?,i Q]T-TiI*Z_ֺZ*@Q6jOw 0Z~n|~Zh~-5|n}sy>LJŇ:ホ4Ii*&OS BOB5䍃^-6V)qҡ )lTi|֤²dlw{Zζ'hYzpZNeu(1вH9l+ˣeB(_:!8n=eL^ꨴO}pJǜ@ԍ ~CdnZؒB1Spv5neiW2mhV:'cͰC㝇{5 $ 8y;jB9N9NTPKu4|8mNbqb(T&(SX9M~LV:o1tHSqJ(Ը'NH㧫"[I!E֢$GI'uј8> *äơYaw 4J{7V20F g"0r+$zSв*Lև}TQҁQ]t;<qD ># W[>--nKE6tO2GZ% ;Ѭl:0 [SA'NTƚt)hJ)ZDNb0pjtrc$GQA~($SmdrG'L/xZr '%jTrK'-I([:&qd#C$z \,[X3-,4l,wR HH.O'5s⟆oIOjLrH10Ve,D^HxSm< _h6OOoJ<8 amyG HDee)zh#?to"O˔> x/ u U xj*̲~ݺ^ʥP֬^3t&;|bh {da3<I4F΅d.!ְl&YDqqđx}muBT' g(kSX$  9n(B Wgɠh׽j0&'mӉ2?mS;`f5jvtjm6@sO x~@@><ѩ"5xT^ tK &(lM =rRQirě)GTrQKM{{μ39|nB#U.j!}:y>^WZ6x=\>&'Sz8ـn"&,VccX:9'r^e [Љ]Xe쥁 HH͗I ``:mIma19@`B6VjeO>̗Q=?laˑӫiYWƂ`~}§c3j˫ȩx@̣p GJD򀲨+ʣJEnMD5>s" MO_Y*/Sλöp@mEWoy^L{o/}S{E>U_?on~K'/r[npv]|Y<||yx77Ow/[Gܡ#}S|ܼETOtA]<;mkFwuw^1rb]zueOm7_ 0ΚkҼU) JR+,F/FQwGN7)~!b%S+Ґ^[O|>l\cr'-z!9cyʂ$]0u` 5e f֟Wd/LYs",4e&yX'm `,L`>2R87q#\3*a3]@o|JxQ=?0ktsxo ƒ]%F33sW˂AvR(uspCRC/EB^a(X \ c â00 W`J~Ud^&^$Yhۖm~v~hӆ2ۊ>zekҬ}ݜK}vwm}vw.{w[P8W B6jSTQ7 EKҀls1ֺ׮,DN>?u:?<>>tz/pO]xˢ]2)ck7龹ؼozCW40)>hۤA!wL׮th Yy SJWYr4TTՅƪZA%%*8 ,f,"#,)=:,.DŽؓO1(?QFNIkXʏN~RjTjGOjQLA_ GQF5a40;ďj>½-/`ɝ-eif{:zV[a[C_LW;ۧ !S:>c(V/͔#Ȕ׌A@[L:hҨܸ JJ&x|r OhLQh˨ N͌^eOMƠ%Ѻ,q\3=hFʺ@uIuc^{@6`dr:̯2_KibS)9v\(7$SskG&ty3nՑ2i@~ǡH $B+YQC$.9 PIB*dSڬ(Ӄ#V L&:pJ*9"c/hN:u$GmpnI ^y3M$4,L-2Ҏu[(7]5em}-d2,.ɬimE=}9MMڸw&Mti6en>-ws,4'BJiIjUPWEUKV #SJQ鲢aճl}ot$^58r+j˙ zXf~>_5o3u}y/2;. Nrh*KoT\:nS,BSՕ`B)% (4"+5@eu*A^U,)ߎ%C=rےB,i[p A GN v#o 3mCm<ьt!d= *cGdiw]nѺp`a iwBܒ45ɐ5.$irb.ާhZ*d:~k)Q5V&oENG,"~ǯ#wSm`Zr,~ib S%c:g5سZH$sAA^:YSku-EEqi9~5wZND<%`TF'P=|P޶x <.rܨFAfp6)WK\T'ŨQ 0ie#O0G0q? {w*Ϲ. 03lgMR#IoaWF mfM?7^Lj iS^&τG>( D7QdDvP ѽltoD&BKXlTtȈ _K^3&*pNʜEFJsA1dNR&xj\C򊾧v[>|u9%Iã%V2q@=nNH钣ڐRm'H' '< Nh`ԄBa9ΨA:8*0%h9hxLs)#p,@% !"ϱWc.P:Vw$ 9mPSsN;ehQ_H:\EMIpTIH{"+ABNzҥҥҥeK/YcKN$Lfu5gJV[sNYɅLDɰΪY%Ѧ(B_ }KHƣTZަ].`LȚ\f |GRR@.XF Er|T -!6=2%d :z|f|Q@sM}d4Tq_vC@S{/%nd!dвơ}AZ )ٱJ۴;wx9 E1}aOebEe c4UQiYE^݃=6ZrMq&XHiJв$ Vq`-hVhaв."x~?-mNaS8=:̳;-+jӲgG~g" LAO32s Qk޾c @x IS#Gw WDrydo*%K#DrpvXղF}k⓳|_F5i({=Iƴ| 9 O 0G Dvapucp|G"й>=̐RK)tF!;K3V?^l27XS^PL LHpMy{Z/Xp $d(9}mnzW : _A`^b5{E "qOʲbL?ePʊ}N4@ohRP:*x>tR;ޜp=1vߑ#1]<6^GюK1iD@~gʏK (0XNB;n>`Q0n\`z:uīPͨ07x{ #QvJ S!ʽOшrS Ix:˷,{5=Ut0Xw :s` Jedb iN*t:R:]yj(3(~p,X1 #.-67]<, 3 gJIKU"Xյ r35SyQVTU %eoȒАK6-O% +IA!WB(Y\kU:%$:1hc#[B0GOc(KsdEa(LWG ~/F֊{>T Ziэ` GNj;5': 92Xm]T.4H ţ2PHAG%BB}'ns)F]Q:}_4V`QL4&TC4bYS$FAˠIZe Mf7qPNxdD$U?wGRSo<$P6gMtt5c(Z =zOF<@oxȆ\ H;$jhNBiRA (԰"娓ZizXI')I*Q- ,GQzpU5QH|h O{H&v~iXNؤ[bX^)WqmOjvkvxdSTk ټP59ԺJMXdg'o jpc(sC -p@rTI'^<ܶml犟{3PuQLԌI6wczy6Mf a 3.l:JVQr `YM?u`4,dZYեJZO3ڶ8ꀾr}sc{qpAiW#VIڽ2xoҙyS2N6zwMݯ$!ZBOmE+lbpb3HOWs،_V?_hw_;I{2Z1m>[=gu{diƟzl6/{~G:_Lvva@ ? |*ӔjHg;gCӬ:~4:`bUR ?lѶjB4A3E`6zZK1xEFaFnsfO8`vbQ޹!R:52 VAIdE`Pв2, HL'0љH3 uHSz:ؽSL%TSIU^dv;Ey˧{J'3Z(禛ud'MbucO_sS6޺>1XڰbiÊ +ݰb fu7ڴnPT(4yHpVU3"/L#*edOӆhFr;>E?[o9ǿEVbd0/908D3%;-YȾHGR$?ږVE!VOw3've[8i3@E~fl`˹iRڗ!B E TYgel_R x:^/OroƓ]K'<0n4]b'olЋq `:эl b͹/cxwF>&[I%6P[jٗvmkp_n?JȎ`2F۟fpvi*F%wĥ]mXavuSwNgz(T8_SZq&n3z"(*H QdSXINp @LMrc R<kWd4H7q_J3"wYAB|Tp٠N:dJvP'4N F<:9srBr0_HBv9f0;X--K'#͡ʑ`X{|yHn ~ LHpy`Kڋ=$6Sۺx%ʹ?}UoڿV> k`͔c)FnOLFa=m%uQ挊*]8rlYmC˝KQ.jӦcM9 巡䤎ߞ΃RNy`:q1SέyM?0q Ij+p rg|[‘3 };"OT0dpP&WԟW %x*xT@<j(- g6;sN}J0Hk5M^uOA3( $uT6Z;7H/\PMQ-ttx,;cጴ :ludh(//$DTO1x01mb+2BB% a'J;b14Ӭg6L8@̟u6Z.:RL@֑_: Z#±w άUmF_#EQ1ZQOK8d$A Q@uN!XGH&cJ,`%Z2,VxB w($4)bOo Ac'uGAsTR:y^<6iY~6OAnet|5 541.9zنnq؆.wxm $tlHgkg앜D[px vH8C5&:#x%՜ߺtx h|$C"yPc:}mqH$]*3A3Tf"XDܻR{bևw9x;N*mk9,uQy`$Q-I&b$[;籫Q C"]~3DNRoK8DɅ9^%<>)T(qGەJ|9кRPLS]"Y4cs ػ %ăD:E*CxNÀTG>t!i'~U^K#^BOIJ^ۖɊ43!ڽVp;Y33ma&uqs/?~Yjߤ(/2?󪔞W*Rjog2T0P!`b3,2k* 9&P z[>:{=yX8|#hA'UrrqâJ8+}Ex˼÷ʇq,f.-aY"ISqi\@Oދ# %.UZ2AGeV1&Aea`xdLzTοOv-۞$p$A}"7LU+ֶTٙgiKBk5Y^]ٿU[q6-Z7]}UV>L×(iݘY$ˋ'ɮޕeXYfe9װmSݓVKOf7ݟέ>7ˋn] nRdqQάәтD"׀0]WWiRl1 ˩n ݋(.? Cm6Q\}F129@LXAlGŵ3F1If((mnm?V`;#'Dr:tecuÏՈ2\۹(* .W=I@8LrSNGq(Wm Z(. P\Gg*6' FԢD<pēlR$ӪgN GFPZ +G'-ޝҗڏ,2^N ӗǀzqm-GkCM;C;PN ::sS?9V'ݷ[,*E* &<xLv`ZF osL^{JAq\:|X1F\$jt@r_RoSh^%!f*աx-GeWJ #4+f<"Qc~b/\4rtUKþ䏏:E*5cFC1 }zkLfHZ#015 Do[& P"*ԺW*U)=o&fJIEd&-9+P(3Rj4ȶyF9'IAL&eBΘꄙ4`=3]™/6DPLʪ>oaRn2eD(L TXK"l 4G:0 |NJbߌ'V.R0A;*&~dWFwX|^Փaub_~|x3V[L1oO,[YG/+ h^6Z>~&MlVN^׿|L?M߹]燛{}e.۬(B-oO^C̅&)%b:9ɏ;njVۧp,H ~0X.j3ݻs[2jqw<䍒4re {;kw c A1X]MfvC%(P("X*AFjk2 zr F`曣tI ,1#t,FOjI.T@6Xܙ : 桨E&2L!S>:!CU= S2奓"NX{.J&$`y  Ĭ  X:hߠNb$R\|G.7>jm汧gUonWa_L[˧υ$= ₯f79&Ga4p?e|fQaj: [pQy`_${HazDJ%%MװF1X(@a>F+M&hw{e2)!<-Q$biSlxe힮ڋ+G%ѱ&W5 vQ_>^ug|sIQE|PI)pga$#Q9 baD2nR+%sR(aそ:@LSHA%2;~ Ce*[Q`4>*);4*' 訰S[<"UX1HqvsQa_KufHWk9"X/v7Y\eLޤeM!2rڎ-݆$#.~ >F0F!5arcta&&Uۅ(]o[\hUY=#IBh@Ԓ%/QrwP8z y%1BFg*6'aTpL҉EiyjI6K6XTA_ ~&MlVN^׿|L?aS߹]燛{}e.۬(JX1/ߞ\Bblf;>x|'?)Z2p,Hb%rV鼝ۢ]^.(Udv޸;E՜j:GNyQ JCxf2( h}{jB9:MD Z}1ɞz{0' *n!7#6QPyC ͇ʣIB~nLh#VGAEY"6.- *z>=4,`6ҹ,X0g2E*te[;֋W:Y`6><= ;^psfzIsAm3'  4ܙ0_P:?w^h¬A?# a]s6kh&֑Hm?g! rsIR =Yi4NnAR },Li3%! SbLv>4WIiHSMɉ)3v0,q3ֈ 45)\q: Ym>8q9* *mRh~u7+,rtqSl<)*,z8aprRT$/. Smv^ZJǼ?q3ovO'W31ob%7Y4yN7Q:ၨe)qaY8<%PStbci(b "d350Oq08' F.- Ck"/,rq7/:JrT2ɢ`tAc18kɨEiH 4CC:xΖo?l0 ^t P}0u&JMi@yʢIZSq3[gbfKRxFϧ& 6T*C]8]Ѹ5RGغlC򊱅-Gz/GAbxYpDw$sYL;yN4$шT祓fʘ{QD<#q8 C;/&MPəZנm*!\v֫!&'>5Es]v쥕5kCO.[2#.3A4cu$5ڪy4Gۋ>Zt9+=? 뚖ϥ8鮼 l(vfŠ&uY/,lz0_&k"X?7]j۷Gߎw vo7ds*ZT%mR9DO? /`PktTlW^ }uU6*}mmV}vCin2OZ?zv6?7߶~yͻfGwiv ʮKTB&Ɠ1%J*r֮ڊK5mEbҖJt P%ʓҋēI(`;8˰-hYP,e eX"[ZVpllE8(Zփ;?Q/6#ٍLe rsiZV ݹJ'7 .jc Vui XƸ.fi躈FDϛ@]Z~nCGW}ë1F|j;%=ԧ^}Gy}YkMǻm't]Um{uåGpmGڥŷw+!wI pSnQXͫ} kf:!/=;kFS|h !)xaJtj!nc~◊&L:&qtr:y++|[w㿵w:"%Z |0)'TEbB# ]5W K|V妾t6~f6PyZ?6h0l|+0m~'CrFj(@rOؖ1STY1}. hCvMI~LoP&o Lo Xl"1DQ#Y0v#XkjP ZjCTHW;m[hPO7>nLA~zDFOkٕ;odgF'MA?5\[?kbwlELY~ׂj:v;?GSWrt5 B\%5tRJڢ·;0C" \É jt3{ڪsh9>qE:t@q!YC 7ۧMø z8xlH?n#)fF\ɮ91$]ӝcknE1 @8t&7]vKA*e :vv>uQlsu\T(Õ׸c[3G#zi5SաM \wÛAV#_lM% ܻUY,U]]!Mc ݏƖhur_8U4(JH_m20CЕ5it@ :ºr%*;Um-]ٴmEUhaEO WC]Ocqu7^ܵہGpkUS`/u RNI6K^ͫ'Nw߰<䥇zK`/ A/</iq pٷs! tCs9w4h&֘t<0ِ4dZCt۩a~WPV&0s6ˍ9 #j&4-tJ ]J:4ݸ S=0NZRсJՎXt͏W˂Bt߄`*7c.gqj882l3#ct"TN:MGSl pdl-H&]3s#trFC3H]4֋:!I%AqG|פ_^)ڬ Զʮ*B~DeWDY+#o0XrŒh +h1Ol7eC{5̉R,HuEc~Rˣzpd2w,&vݸ k,3ڒ]c0h> ]=0Yɐ; kY:ukܘI[ vB3p)I? I;&Qbv$ڸ3%kd9|fMaDƊq*b 47h 2y! } FH AH5uGT٪ skPW%3-us]*;EY~! #ʓ4g:wOVɕ=hāSNOu,ۈ.;l[J*`'`u׃oø~Fq=>/Z wrz(B* C?>Q+!SHz,au:ŹW=ll*0%1SI1GÔp&)hbs `%yWӂTv~eerB]ѵ /2פ,=M>q{9աMFZaI-CD񘣰*prQ8(S9J옣AtW0htr1(BQ7<,aIayبT-GPYqTQ:Y͘NpT[G#v K|Is?{ɍ"`&`|:=̥|{/K~eVRDRa-+;32Dž%g3B =)v><t:s4. <Z=S&Ϳ;5+굙.E|Ww:zH~rM]}cv/}:z¨9u/0@ Iڌ@{w5@}8u&J]/7 C9 @oCh PEr* ж݌(.. 1q8 (`]C tdcAXZBnWeSӪ ]Tٙz  So8$mDO*1d j<9F$:mIXuôB -ZeLDq[ZYE%eT[,+S@ZVP=:eEh:ըK˚<^d pvpEMɚ .*,["Ww7yXPDFʪ[Wsk?tS<ڝ@ˍہ:B[~;ՠgvu뇯_OazJ^]Ey{{f|MqcAmJ՗OpKpt՛jy7?~{|{Om_nRXUJ̊Z_tE׷-kg:!\K-)-13͈ˆLt;< 7L&;4V',|&4"lh~-C`lTr G+lX,KXqj?94g_YI:X\N cSNGQ/qvv԰0OK|Y3nLg'."Dp08TG9w%hR]E"߶pFt8ҤY-Uıtd&׏]ĖʩGeժԪ^AbUxu!ww쥤B>7-Z : ]hN㞘`/gdBsPɜ$OaRJg΀ h!UQaA. #ghxcWōiB։B`3}F%j^&Ur/p]f1&9O1b˓}L3o+5*P@r05w#%`v!6L9 4f/쎑Ouz !)ek.%n T @\j[A-vw3m F* DJJJiLky-tx .²e--q  mm -Iyi׽oø: r(p|a\]G݇q!Z|q0.しq0.OU$6+4 +Yh¸\T.kxY6KV{}۬ S" l fVؐ6~<Ƶ(/̨_P f_ns+3]fFQ9e._- ta 6ڿ?prK(ō1JT[xwU|X1j08\t*⹌6Ug=Аμg'2&RtY!&-Ne%!,e,w&6+NJؚK! pl"JFDEҐ JURj(Ӡ* % 0#ʓκWItu%T-ۚhGZWEj4 I'[-Ӑ}{-|eLCDԑupxc7jgb`HƧ/E@Fk1#^C jUyG080jPڅ!Moղv>8J / cJz(%d.α& 1t*6icTX2 LtKԈQK9zXi5l4K&$'мX%$N+s\8(luz8(J-0'fX$So>@XH"Sۧy،S2Rf2ߔNM1:i`7+$jZNN5'FA)6 .CXYCmMN>JP4s,3/#ob_fu%i×Ifu]17H]Xܡ79tQxR㉈ʏ}6EV#gڤK0aw!c@<9k?扽ԛ|NxܐzCa<ᴙEۛ2]rnF-В f`RXOMv>Ih-ҕ]Y2wjoez h^**Sْj\ABŝ;-8_I::Ne hȝ'A1o+E<-qťn~5W:%r(s.C:TujP.uR03:Q\h_<`J.?{u|(G*ҔCX[sZy1N.e% rtK Tgrlbh^"@Z] FH b֔ZQ +2$Ԯ_RV> #ΓïƓi5B)jWt[s,h,n~9SKqlJW x2XrtÔB1#|)G(;qD"k6&Q#K}r~vVFʪ[Wsk_}U^[լy'~rv#o{777^[_ ]Wq;z{ I%ˮTTxv=ۇW?:ٶ`\}i+W7&dw?|Zsy}>߾SۑTb&(Vv]]KVr _^߷gyrҗLG$qLG,D6-kɂ"Z1 OxX4Vme4Yn[~U[͊ 54kQұXNABӰ+6_=LT8HtMTQˠ6*kx qmNa3&q(JL:( J0,{ApWB&qE餕tJ0րl'GQ%G2f^y&Y}$h:%9.~kV7P25^zWlN ۆ?.7\_鰫D۽m2c gf :H~6^)ULH1"Pw8lU^x4*"I},AvnFβ2تrVdnC&2$UuC+SnUr . xk^&SCwm,UT N:pƴ0y"杖8r_ _ 7^pr欱 <-9kbJqOH &Ҟ$=tNK: D]~XvX -LT;﹇A'%Qwrrh[;FF5yR7CHS f.4Mbʦҕum]4uURU* nDtNZSԮaIz'.1kX#)eF -q.C/;EH9>AO j]zG(1'v]x tEi*}G+!)HG+1&h#fO<~,?6Q~~&K C [CC3YV(ah ]n˷LǦl !ӱi"JH26ASo>Tles=E98', >U)Vit]Z0{jtv~،S1;I/\<:H\V65.+B'Fe$i8ÔN90L19KY6r}MJ935KθJsPӐW?gea)gQz2T|`_]X}](;{|F0X  $ѨPRs~oc?x7O%S"7$D5 G9\(hSER]l_݌(:ʂg4T1+ xoczq$FQS3K`),(tQGP6eB>0tSV:\M;!u3H#xug kTQg%NL{x <-E @f h1Ifp=Pt9B'H~˸^F'Q.=`Q9&9/hVdyyѹ)'Ў.*bҴNd@;tk- t|'N @u6,Z._Mkk1-gk.#n0zjUk]ҀM[Bض֖ Tmb hIqtڼO*(Aڪ j(3Kecu]Jw@iŖVZ9Iw2{y ;$} sLNK-Z4FQ{~ޫ8繛 9>tχ7CS&omEH]bEfpGAXE$>|}5|tz--ILLP-ә:d㪾@8hpU'Vsg+rn9/~Q÷nXADo>~>,ھx{7`EܚuٽS|]hvlZ;::hm~U^Hpη[/Dyo:E.2]Rf iĵ-啵W-8gI+Rʘ'[c_;4"6}u؟Ɠ@}ڮ G1Щ(5B׉0!)uIH#6̮{gv Zs讯IV1NE'-ٓk%n:R,=*I[ݕԺ%=r3\;讥xJޗ+GمIV2 讣JɃP5dm5|C\c_jBMMZ<4kVJ;) /Q F܅]m5m> %]yY:+muHz+K-Az|tW9VCwekЛ:<]\f,Q5TYDwSmatWNb\7+Ng@jQTNtX+ 1 )?Mtm?xqOêα%=kz`U0GNWuPq*T~ 5phqc (}3 5pՋ} ރq+wdLq p23Y"-Ӹ@U)@uIj؁xh7Ӌe05r͚۟>2٠Z#(qXw]΢j5^(=lBwkQ1hFL@yK 2|d$#ĘĈA uM n*S UByMivɓ4 C.}xG>}CС ou=H}PEXI 0kܔ%9}Lu7Q2bF'uUYE5֋9d qTq0!-#n0|е亮5hMJ!8- wmW^Y `F'YvustZ; Ӂ;.GnPwbsM]SCoiH=NC,tχ7C|!^x"E fpGAھ-_o'X}bGR1"av%uyfW ܈ufcCfײ!75 ܺ6tVa@9MK0y&шE]sM6Ԉ7Kώ縚EÊѭ4[~e5uF:uQf0b( \KxL]is{BFÅ]*cv̮r-TA9f+uxXٕ(9 M3YP\)$&ʉ$ *O'3(UY+(ӒH,PgDrp3WǔX t)ن]71#/@-*:w_9nK|jNxUfm23 gh}>X7ir*o,=_N `2H"Ј7 [UG9x2fcj(?oAnf̓v#lUю,m3Xކݟ>25an)#`Wͮo}lhwx1z7vysx;gA( ӹ$H@'[9;yV K&yK2&ei$pt@'&5p FdVs)&5H=ˎ%mk)pNXn렎6c.}46&]AP[.;"S:#5͒b,+cX7y:UyFY`MrV/mǏKy; /΢-:0HS; 6xސ&E Qih{yzb C `3<)=fǍ}.pC+`6`z5m]a:rhv4Ó>jYu̮B6J;y9k+EwkbO]ht7tJqgٻ؏T́"%GzvI@r]kxn=m3$jƭ@l m]/ u]4ki56VKM:2vo 4\[ Z;zt(Mut'T߼&V[QQ j(5SE)[,BFLBJa.F0b}4qout%!QqWVS 'k˶^*Σ.F C*Ft<l: X3jqqfuM>4Ҙ9$uu9_6Q֪Ȟ 6@jJk=r p`@un@@Be%!Mz8ov:jZNfFbDaORsL\7dp&F4R~izppӖts5a˦tZ&] AZJ3W7S_E|V+0*gˑV8l=}%8,A-L@Sc錼rUyyF^Z#oz;<]\f,PT+֕ -r- @) MӡxZVjisbk0'3dsav qrydnNuU0oS~gF5ppw`*PՋ}60vD+…w?ޙw5 ˍ}29t2Xgtm D;G5d XF8ZJc@3[rspt:uƬ$Nş@bVa<9GKv?Loli-T/d2Qe>#uHm>Z5WkL];QLxA۠z8k9Yo)5ɕRKP> rADf]p H ,cOpcޅ,kX;QquU{%Ʊ, C:dl8f, N6z's\N]Y357[kBU#knµFWS'3KeQkt+45W_Wʧ*G֣ÐTG.CSYWzsʩ8'C-?"LǓ?x܌?9`ܟr-ɠ*(ddIO&G6t\lpMע-X!/)9v$YR4ErJ=.wAXfIΗkk +VAatINʪ[ng^O׷}/Yݨn7[^=w77WWmv-ďֹ-)_8;ez]?.vA;w`vzvfyUsuwu'OmBۿV%xPŲBj~t%R5(]{_tOH{)RZU~gYK/VZ`F^0jU0Ԫ,ˏ]͈f81m~l?%h+anVDmwR&p NKZbixYXUmmicNtRqna9XP`UO1'JS5`6H#$̉|A 9UyT:Ygd؞2-X@N94X%?+5rx'0oh{10N9N rf/ i Z_i,٤Jt/+ kgQe0IZf*[3l)(l?eVv٠w.)dhWP;iᥫ_yFhaR<O9hs 3`<+lN%d=t̮g8+j3bԲ)kW8ٵ~:1+BQ(J̮Mß 5-)F'#94NCe#FdHLPj/lG2\|9qDaw \&9-7|A#(w])0qh8),XZvSe77|!ſ}=CZחEO[ Ea(< LʢI=I ea Q~!\j WY-ű[j!4@=edï߹ (nh6H ܱ)!r 8sT:9e0g8TF1KH  "O# /DazK00cdhuaQ"*#bIn^ guHV\\'>l(JD;HGD&+yEVeCK;ṦWN<䄂:ZUL 0 G'r@A?P}hiO9:O5#lLw9҂|b 0%el7C>FJYS5Pժ,K 㪦rUj uqJ XC eIۻ-Y҅kU!E궭-R:UvEٶ6iZR\U`K^4mАqv/ 4$1 ɋ¢1E.۪ ѐg]/ x;u6`R[=N=!E:Q1 i]^jʈ̹$H S" N%! Sb5) i|!J@a619eGv3iw/SM\]+dP%ЬPrtN3p d(JH]4 `W%S`BҐ4]hH1Be, )jӐbtr֤DE4tBMJCZri )b!EF(`NL=eԇ6<+eF1f ESjѓ#M88ˤ\EEmV3=( it˜!7CkJ<60 Цԣ7Mۺĉ@İ@,rh69nfD31b PT;`@J[NAi-3|@X|ئ5 aEY-[ʷ0(fo&9#=Q@ 8.ڬ3A*<.|q̙ <2͗9RIRG]0y"}8̊:/Wjt|VT&@tR8ȅjr>60J*'4y!`2z  9ڪ"퉙@a'G<3ԣ\k d:W&d܉q!b3ݕׁ0 =DW%M|*7칊,`:T'UfC=z-gO};m}ɖ:X]CVw|X-iZ}f^t.zѹEEoa1:keM֕X Y*45"]+]{R?ʮZ-͢ n ۮ7Ҽ7~]]='vim5lcJxR{ZQFXu ڊHQ%6h*D5ں`EE1ρ*q$f,9G$933-kkݴ~V.@8-Ҳ:Q}xIOV֬!E*㎗bc5v$\Zetc^ZV .=Q1-k=iM+3'#]IiY؈*3ӴSвV4 J2Z v|ɝ9qƮghjyȔ`a YUt^QK9ZXФf xMX=pzM\Ą{:dұ Jӳ:B(=$g⣌Ufh/'~LQ>S mN>nTE YHQK-'s DIv>WlVu96п߅d PII&҃b׵j@Da{H ^oJ10+*?opѣ}Z7@M) k*]Tlk+ȊUP]5*.F#ж~Z >]߾f]wM~W?u}q͍Uݫ!gW-zw //on\|[]vU(jNqyݶ Ew`vz7˫^nS_]bY!Ko.%R5(]~{| )tD/IV }ɊwOViSk+ ґ>տݳ-aR]dL8-F=0R]w4.&i-nߌhjg'EU΂TQiIi*xVl`%:_-UK!'ل3lXF # "޻:|$H=e|sr;zdg (,p\! 5 :p\QJcru؀#8"qȆvvu~"rB=@(\"Jy3X%xړN mҞLƫ!V'՝ >=3c+R>cŸvtxk80<3Ag+ρİUfG:-BjSǰeѪVMc]:T *V]T4ik3^ZRV̷c4b*:Fi7np*[j[l5ݝԆ0mwø/q=<ø(~Fr+Z@P\ҳ⎗Z勳.ӿq +ȩ糎((c&͈<*~PTۦ,.O>15Hr@*` RY9j$_>~-I]b[ӲDJc*MerOm؛eқe$U-,{ fUzey*0NtV"N IE~DCU9:^4/[zOc FC;o@,l1Qa 0jeΉvDe zh%i a?-8 }.QLj6) }l-0*ʆR9t=l6~w]Έ=,*GKiC s4t"̐B^LQyhYZ1{'_fhjqʦc3`rXv?f̭ʪ(Bk6JC lX8Hat#y_$ڞ2=K Eǥ K'?y ⨐G$1㢟:T9TB&ӎ :)^ A@' D>/>E΋9QJ8ŐsmU1d椟_YwQɁ~V>Rf梟Ԃ{W|*r?RGW~|D3zt/w1})|Y\\ܵk;e4qS_~oۢ_w[A霯eo<"m?<%X? ?Z-ﴝm۴%Q5du?.Kߒ~o.zѹE^]t<^{m|J4vIʶIiUDtm.K!,~oԟoooo)؂,̠_>*_o>߬Ku˦WwwŸإJ3K2HH *%mmSUҶƢ!!TT5DжxcC&-{3,3Z$퇹9oäɈ*1)=rl1fX [_$cr"Xr^Uf7vc6GS[% gnldiחYJFiu`@qφȰypzzd\:zdNLFHG#7O'Wqw5GF((ILH~(DR Iv*alsm3M I 葳.?EB(=.vps/sqWiL۹0X}}n[jo,/KRwZ**y9$FQi/G)$G5i=74"m626~+3 1+B9֪llCv^To!U3 KnT .5jx)SWp6΁j }:kxJسtg *5R鄗װi~nX-ZApc}1VkO|\B);Nj.Hks k}~վ)Y+p;/B+Y$bɼ{_ lAտr@U鈚-c QsU^,F jFcDB?f D,~κgS`e͂QhsIzI(z<}Vd+/Jrub7-g+?Uk}7*^dO98zuQ?Xo)uO8Ar~Z:' R#1:lŠ8 (kfFU?`3r4n1=8ϲ0#'zOy3Jn:A[7둹.w:~|_>[ԋMY~~=y|YݗO_Ga?\=MDMu~үho ss{%gSٲ}}13' Ԋbr8`fC,͊:G`M "3C,-5̾AYm6׾nؗC-5pbJl z8j1JٟoǛShxM4Ę"\!"~:y_Q&ɎL'ؑQY" ͚NѬ1:Ys^= .Sd:3aG o8" `iJfcH1 r.Qmh g~Q00qZ椧! Ln,=24a4 Q1Bϛ32V.?uUy602x$%Jc fQ%4l^h̀jXʘSh)Ƙ+Z/DϋL ^ELM^+@(,nfa1+˺tevwsWu`5`UV;'%^ABŢ NJR/O '~I"_n6OImh:-}],÷~:SN ۴dNWEr (ُSGc>lũA"b@!D&I!m" )BZ|'AS z:g):"a4s`J^sa ydŗ꩎R~AʇW7 ̍TbdI[mdIH 'c-J['*촩 ҲnT Z6eTFjSN;Ԓ(Km^%M*6BlJB[Xh+ 'm %F;[!Djz`ʶu pdQ"V Dx9Ru>qKUÝ`HlH07ZuSARuca$U P쨴}pLfpN-㦄W%k>7%ʐ,Ņ [?/7[P"xIݎ*MA`՜ɍɑ5tC;KS·8o:G* ]}24tp~]>OpN'QQi)!8,j#bb0o80Nˏg+c]`f{1bv+j *J mdPNd2"!] eh̩? b॓  iPG=7&6{[)?[Gqh!G5st38; po?}۳Ae=ݯînV |:2]> @P9+/2:(@slF@Uy@ȣpLQ1(*Lz7aq~0 ['(`:QDISi`Ws?)+N4CӶRT3M5B,R-Lϻ3++vmB>7zoOLNc | g4eM + =-HwuNK'1)H'&|INaD9<%uX1[} ?CKB&n^ :A#D؈|HUxh@ wY)XyY iH>Hwܮ;PZTF!]&bN:4¯:{(˙$W4IP"۹aC2EަeZ`-{u#K*(YiU`Q+$J7 0#ƒBK- e}cKEaPhۗ!xs6RK7l|wٟڲL;oWo5zue̓q|ߚcMr=|ݿHNs40'_B k|!9Q0';)_)>.Y18q^yG2QܧS/Oa;+>lЌ ӶcOQTq&mOuJFQ=B`esVl%:}Sd(>bL%EQ -%ȑ.3t{*lb'dRҮ/]NLIǝYAb2[B9΋ 8hW:C +$J 7ET$ms=[ΰhx~<UQT:((0%w=(R(L]` M{1F0(jX٬IIt(JDF}Iv5Bd9KؿFD{αR+WW#Y'6 M ԕ**]z f0V%XUE.*"ɵ,(O;.=n7Uٵh7FZ,|Y۟grpUXvkV[oWe5C^OEnO>ŏb{[Q}sSSb齧j-c|ի=?֟~~=?zU)Aʀ Yښ㪩Y ; G}(ӥi /"}(#CtjCbw_jMC’80Pl }bn\̍<(E@|'CJm=hLM z.d0p_;G:Сl˫s{ \dL0k,JE0pYP.` \Unbf@5XN)d0 \T H=>+T_>KG[ֵV3! fNL+A›_|#2ʚ# 2G{H|y;AJ]:1u@e| (F k$QKU2c[8qıp w?!ERhk[%Bd«;o?}vZ5 +CkEZ:gw+y6X#'ݣHh=JY_>$ԪV'oZK/G⦀]ڹ4>1C2:x(˿6i)7:ƾuf^V _+^Ǩ ~*K73Yi *eU,b|_>[QcyOevߗUekU6R;b߅ʪPuK [7r$+]8[Raz]*F 09'8P_.A8P ) u-,P^ 6{ԝG+dW9+ *ߴ~[E.Ңg;op Bvlj{|z-vW7cӂP:92odlrC0ĵՃ72 7672>t#c!"Ԓ$ oU_h&PQ3K2kD^J1 Ft|?qWk{kO 4>ZwKvvc/ 紬iYֿ)Rte-v騅Cbe,]mv%٘WcIYSTֺ*o**ޔDA+U*\ӔX8nIYɪe Z7)A],OZ$ =ޭRjEU 'm6J :6on( kNkkj`ڎ/5BATCB{BB?XP6dאkħ$(MTPkTt!f]j\釿n/&=(Buf8=Uaώj1% ~!eٖRYBUΖhF(XhP+j,5۹Q\zcE@%N 5gՌ`cq.ҌLK [~C[̅;h;2~"iyEfTI|/2A_kX R|vCN7l2KRh ghj M ;G{ѴU ts49ioϝܜ|=>wJSFč8 >ҩ6~ȈѣK\%맓cED(ʖe6FHع,"t޿Љ"-“"$jܤ1:.p4I1r&eIpXedzv~Q '4if]G5 iܙ:-"Jьl]F<۳=|Ѐu4;kEx-& tc8G۞W8|-W(*}wgmq̗3 pw_nf4 ܭyvy*!?I|ĐJ0aԤ^fJR^%44 )|7rwUZ吽_ ?mI4鑻Gb+ yZoG6XFߺN|"H}ߴ3 - *.]Ѫ?<ѫ%]BdsB ׈GA6aMԦ~sF8=jl%~0DrQɽ﷕AήҼYE[}N LB3^>]WSib'/6ѕ{c2:`VU2 Uc ϸG򹉨eB9v **:MoC.)4 4q 5Q:R=1R)Xiv QU0mlPX)F.9e ]q:BA"_!o eJK37*K&[,.?#c-D'ۍ֠fj^Xb#Ss{a+>vAe;]e;]vnb:z 0cV*"^3=jV.z+>w@+%é'ЖؕzqgtfA۔@gUS7)m[Fb!EbD@7,3`( fbŖ8v6aHm$*GQ^q4(uݚ)mSRHSl䉍3>&3!մ &kh(IZz3SW+H&S˜"]nͲu/uRM,ҵ,G4*啣kcjF. WL/;Q4RPA2zr4Q(nWHWfY{$8ʅJ4q|IHAw:U͇Mrz\8֍z/Fv= B_E܃#,V/0T&͎ƒHdû&kgX3jHbtckߌ-'4-}o0eUѵux!&֊F(/+i2V?JyD_n 3>#onI`xƋ XwRyRDžݪ# ~yGߝrSqc^I6<"$Hםl4S7J\Xl- -G┡#Xt, 1rʽzД+8im>M0 g-Æ[?=32ciMDc}ŭ2TY9*?T\e>_ݭrXXr|VqJg;/| +nddJ,~lvl%F!Ԁi]kgцT3 cIKTDr]'!CZeY@o)FdCeTKL#mχޅZ3a1[]v# +%koK@5;~:oqy{CK)hu=o6~ӎlIlݿLZt)uq̩-VxC Pj:̍휉/E~0 oDCo䶩5sTTz>-рv7c?l:d*2œU@@~5FXA͑RHB#-$4ē[)jr}VC^ܾI=)C}|1—vLNȑ޲ qr!>W3y;Zkïm sNT B_(ps{A|DV7G }ۿrIgm{ ?5=C((O!Voz޽7EnO`ӳ_k`\OY  {oI*D6;sɝL /npzNIK]=BL5^)J6| MS(c-E--XM_ n{ZKo(a#-9kq挧 :I)Ccckuɛj ,:<6Zth CԺ>8ex]`U:B:@Æ8]ZjkZaiiwVV[UPʋ@tyMa-k 4i4L ~-[P˖t'׵{Gc]32ƸLPstXM{'̲^: Zn7펐g ":p5OR#$XM;8pn,uȷ6J]qOP(Kʙ "=( vq$|+h9.c]lX|)F%&#x7aHRBv 9.p{*JR JTԡlE1cKe֔;kՉk켝f?yꌦʇa)t{2cPGozIK}kӪmsJ\+X]%U>wVrJ-~Ut2c2R3]r./t9N8C]kAEղ3 aӥ]~0/5KLڰ&]<`Z`t |O~0E˔bVh%sV`F&.N^1gu:`2'R1 ˘2Cmhm[>.[|ޖMV@ Ugp*1403`3Vr3]ϵa&{}/so#cz23,lfpI7s,qds&*9\9omPj}Bl_hPSVveu5a Ǽ xm'Q2p\-su%*sӝUA,jI&1-C5t:pQT<O0Unv\4~k8io5>qnA_jVv_X YuҶPnkvNJ!,^7AB1զX, 6Gl\\I>?z-hn60Ziؤ\3eۣPqahu%}pT X(}{ةJyT_bžLo|jg2zKah-W, wHVSc)cWQʰ/ZpRg3hj~b-kt1W KIdiDxQrfUo IQ/~H q`qń⠞0Y EyM{o*Q`ɴ -q>qRzB^:$8t=a~ c:rJe{< 9_ޤ/:sʳsD%qaZ`MTޙ#HSL.ㄆ a z݉6Aݲ^5g-VNDvJd%/@jyxgg-z9ZtT 9^we$^ r ]\Bq9uZ~Dp9<2} uV4-Ô;:h<^l,hFbX͖1S|Mt0t(a HHXiAA#E4(Yz0#Ǔ1b<bhKn[)-T-}uŎ'1ѓxYDuiZӸRwUo[e7 p2Jx:M m]:853A N!&񭬦۪#+kF[6޹5q#{r%[ݩK|OʧvԾk+,&8~4H 4J"Q3iw4~j^?RQSj1Wo)z^mkWe_m]f/.zl=[UM3 ؾу>ə,ʏ}zt]>"E^XW BP J;P*|7v k ;RbFA-y|޻;Z rAvQK'}shJ%`?1BrTT>P`Zmx Fu~C]K(-MY$= M]~U/uNҭRu]Kѵ0bpq+AW+ ́%1lus|R㡖jR<ՆG-ZF-ma3B-(JO D-SmhԒNAXq=FgxZ 3,%:"aQK:9G-()!aIBc1 8Mɛ>J)J:3;sCquT_ Z dӿ>4wtC!&y!%CG'}+ ǢP\L5C Pi5lPT`܂?(HktL6ͨ7g*$6asM @)hm`mef}I sȒUF kXEQ^؆[ؚ!lI+-$eLQ!@aOL %N#T'&K1g3|t"H0lJ=-q^r0_#Ցw|_)zTMAM:H ri(dTg8%^\@uibz4Zs=Bx0NLYOX58SigmTќ䜩3\i Q˘CۏRO :뾓̙ 2:cCʓg2hɵ6 X/74h2QihyT\"UU(΍k҂Je1byKRQRKfi ɝXS$JW3%2%*F)+ B,LAnwU'=)۟?wa!wVYqAZmYySb㚿ۯO{ 5zFﯗO^߯o㫹͏z]YfY}Z?l%>X_Uh;w>d뇿l!>߹zC9I$ Oߞݾj|%{(?tϴٔ|2/GqYxOڍ]E!@uU. i]xL+Rq!S>G H$B:G 1Y** ɚQH F ჌>h t \ rjǿnv9!L-2bUyKACFuZIHѭ 1Z`=,Wu*UZUa*ԓ:aH]FtJC '!Yf  Q$Ģ(>d"5fOaaf:SJȔS?Y4FȦ1z"M!hr18)L2> (xiTH>O.p" dh R^#bb6dtPt|Lр6bt!h@aKș0gw%"6 ($ 1YH(I)ЀŒ09 (`FѝΤ̰L B;"cX.x9לٟ74^l-k tϒl/r6ƙxIaGid^dÎH̨ bG I yZ +`N6RGH'O JIλ"sjHq޵= 61̫(Mz:R]N uWvVt84 Qzcshκ'fD32'М]\LtSY1tKuTn]WPKF Z!C4s2$iw)C&QY(jF@e52HBidƟ3B<|?RL-t֑X L^@QԌ 5׀B^J>.rbַPHP*0 u? ɶ(5"X-p!mCQÀ)_'LE!z{Λ֔N^oӞkރKFmWKnr_ Ȼ{W\j3fUε0pCt#:}I6*@64D3u/^>*yP\x uV09%na2*wDq!(ƀĐPHiAi5~W[/6,_'N `D60Wó{h CÇ_A"@k2-t##%$ZmM ind0LO 0p!\0.Jw]j0^y EBN]:=TҔ;Q TPw+#j,(|7f7yyi\jP~=lYwO$!|H"cQRwR>͇NȲdAg $ CX# 1)S6r [U%.h O*Oh˲kZ "P53dD+Um1*[d5㬇'F7 7p@gCSyxC7:Q8Vf{f; *㡇im>dJioKUv;]뢣 Tk]n'X%mU*y$Cx\nnɚIr{L $] >|_1^`%I+oRv]KҵƤ출[%)%j'u`87oL3ustn`s%FUx P0NFǣu7z6wGWB\  6,ēmXBNNI^īx@ M2.Mb@\5,P.D'2MtIS$bpWԉ1C@Ӏ:N F6;:|eL'p)BIt8%"2<x 6*t 8J#u7+$I lM X>  1X4,jgp}g;I`"BE_PUTR%鸹bkU]+,V0e5/B#8Q/:l;1bc|/¶V@SԢӻ|:/NQP<)h' C\!"o.eJ3- Iq WGr >V ";*2qt=!J]g85뼛 av]R*$%4N `bfי$ #3.3$;hOu;e`_0B>O8O<`ZBdSX,7&%JؘK)> PYQgU^ԠZ+Yj(MQ52 RULnp&yeπAԊ7,J,2uʒXyd*,3ZrsSf}XfJOQhphZԙ0(s;LV,0J~﷢|p$Z) I ;f)/a$F#OW|P^2NoPzۍF:PxJ!IT?Y@HBd@ Mx 00%\L)axߋ#.JV YLg !T] a\#soZy]0% &q\뜖NoXy\* kT0181F,xJU(Y^j8%͎+UYky>alNݓ}gKC0[BG[.-ˤJ-1#$FkLT>rɨN kH7 80nEܧۂEn5L%CZ:Ӛ_)k]R뻔ɀZp`e9:*LEBXVSmp`UZr{817kތ3a3bl(J@(#A: M$S)XòtBQb@]DE C :Θb|Az T֘NcB3F0㤢#GWf\fBVȢ`wA`?%^HHd?9Y,kM֟EI<'LQ5*Zdzϛ]=-hX8:շ#s!3k9W MH›XǤQE݇#bDIݏFG=TG9FC2MP ,f;Efϊ(4*itqx_d?ck*MKzY8˼(J*MS6^7Jw^' ' RXF9"ڑo:)_QRqWNL#2qa5y/H'$[4:2l]LIT:y;;癟-}ADﲃӊb%B:a-BxUPMj; \9.S lfFJ'Q];@uRhq*Q F3z*°RX*(aDq"T8Q\]7miMvQ\ }!ҖXCW&r@tc+j k}0FUP*5CP\C?>L`6K`%%%l.M%dA(ehv\*E֥ "3וXj!BioJgI6ƒC[KȝPM7u[vޕh/V/IfUu ŵi?-Ku?O-gzzXqO_?~E|{*wӝV_'WvI%S.}h> -Qd}M"VI :31:YDNNJ[K 5.+B'+tDKp x1=t)eTQ4$'GoLjokN G2Q<$ڂ 8#P`KKᗸOMυHXDQZaD|9~QfDeZ$c?_+"Y1&,2 !4%N,h?XceQ%EyhMP @Ԋ+87,4a~bA'29h.i'~PѠlEy]nmQTȧ^ړ` w"gJQ<1WOLO{AREܬ z\_=)Η6El=!uŹ)".^V"yԭ/&!I:gUT"LYds3+a%gT¨u> v;K,ρv*Z1`fؙ/-! TG.%"T>dw^ E%*BR ƦȻ瑅%EL),m!+-kU(ɺVtҀvngIɊ}, ե6ΐ *PFhPv~ 0,߉s̑/!h } s@8DVm{ D"¥U(uxV6ex }q0GAj7uXtsg"u.h=034\b)1-:-谛^d`bJ8/Ơx!~1fBTvve|hQr4ʑj4t:KSkΉ&CӂP&o- MBt9 [CmIA'Et8~w<(za"Trs mNb9QTBYSmdQN,8>'n9J(Ԩ94sN 0G1YE,ӓ|b񑄎(hsrl/'50ģ.a'5֓ڂErXsS*-"TPK rKp% ڦ$SIҬtY-JW5r1+,&ZOլˎ=IڤA[}_Vtf-{{~I94k40z8=*JE-l6Ծ971|1jYB5'Ir3s_⽼` 'lR mN޿c,_V ef~K$/jcbtR9EǺ="N)(r9Ҙ iJmUz_w 'Guůp&ѩvI'5VPz"B'h[ V 19G28*/UL#>Կܵ X"ocvO/뻽J < $ +աOQOeQ%٥3Ll_MD1zFҌ>׊GBY(c!U*PKU 7vdgO7SDOӾ` 6`ZE[@o#H׾ernLev^:QBس -^"d i'2Vˠih^n~G6_2"eUE؍04~HWA:ŝ߳F}g ]E# oH+.-,B"uMZ6naKU.Z(L җRk?V"*ZCD_L%U8V2lݺ | $AØ`ky zNL'2iC;-qP|aaDNH)5y5E_njlA (:9cZfB?wtWD>`/3V?^s].J+m1noU{?{0vɦo?QԶ+%DcTPdm!J]XTZvued~-q4j,9t`Ia;IIDm ˓WI IiиD$ڰ:V}ԥշ0^H +~Q~lLFnqqi`zWZfV2s: W%mCH 'V%Q5:VH(q\,^ .z5ΛѕbFR#RcS)SEiY6N} kAO[il )K( m@Uc\oR%Jh^)k9l\7˵Qyk4~8I».?!rMMRߝxu'oDI+Tq#F#[x(UL'b*yv.N=oOaQ% yK |{WCJ_<ʻD X(pվ=UzBFDX|/pk`RMvAR NĐo ⠂~=o ('&9LX78F|Ь:&U@-wf'tMΑ% ȪQV6uSWVs umȔE!زRj&eEot<)gOU`&6֯9 5|0l%|֪v~5٦sfZIR@FWjFԦRE] ITXKR:ǮD 2\".wA_˪p#|\b)/RX|Um.Yy%*1Y˲m4{͞.smUb1gKts ZiNvsH0s3woAjT9 ,d:Q;V 4Uo$a_.2UEK 9 ~B7y+x-5ڵqߏ=3[?؜<%jw%Ed ZWuk@:p1{ Jk֢j0| S𥹯ٺDClzGڔUbhVa qUZPisd+mkk jY)pl[fZ7ʀ7<VFKm0{Xqeo*afxA2oKN9&)+F3P#x-[|[Aak#gPpֵp^S- ek"4JEUP ) )+DVì vL,rVOܘ{FMw&ZOZu;[㴶U]q hϵu#I$I%%jXΑM &Ay]DpUptM-BhpUC+j&QK؅olw] IS=>pg.Ζмma*F%H23㱗> E#w'YШVR[JSڃ A|0ÄA[񦩜1'g_Ccߝb] KjǏ2xM޻7vXv>ݛ+K}>X6u{۷}z߽"ZyE?}g]{7G_Gm>uvO!wrν_S+˄+ qlp`;ÿ~4|PoJm toOj}0Qi;=`&&qOϲNxX[M))0ERYR9KYS=*EVuVȿ3e c4ebGga+2%qniZ!`ZvP K4U"mr`TS1yQgg'( M>fogSmrfoZ1{mήbLMPTJZETٛz3{3 5)7I'Җ(fo>f`fR} BҸiFF.(j1-@HZ蛾0@f'[jT֩sY \8ZY)hV$,1S%nʨ &wK_wƼ3)Hc3;ٞKPٕ{ py 4W EPBeT, 9xoFT*IFpiMJXsTzPd% : f]?+ޮ5jU{%W5w١Z_'~;+vb>5θ:X*ތj sҩތ(yC ?rͳbW'*%V" Ҳ`hjK$8 Qлv;umQK׉Bh p&P%BV:5H9|ݱ.1NR/<{^W@oG+o r ˴o (F]Ag+b ,Ufsw6A7.qUGG<>l'۫6eȾwR{ǧտ߽&u;͇=\աҭ{8UֵWN^"vע6uR7!.DM]׆l[CaUK4x FI^ȖeMügێL*鶪?涹.inul޷?obVI6Y){?ro˻RVvĤ=㋼W[W4A;ry΃\=Hkk\dck1D5Ns8-ZlZ2V67Ī%Q_翶:ۖT5'_Wc?k6W9ç.~-8x5ɇm[M3"45Z3M)&f\5uh 1Vڶu5QaL uUekPgKRKmc,i(ӥq{;۞ 8I=\zN"X:\U umo֊I`5E" ocxE\XGtcf%幺={$Su׶ s"2$y1fGcؔ"[dc628˗Ffd ]D+t,'+gpf"uo5MNR*Ti_JM|^R%"#`hj"ftYܜjOz8?,IEC",͗c6=,^-%2lx1=]Ř cR$Ua^4cO1f):0fbDc%fZ̈e_t vbP"'K"݆Uf56 ^듳ZԕY `k~s^:7Y1fFUآ7"/a$ۉ1fqqҗ0f1o81B@ia޳zgyޮۚcJ˜Q=hYUb]ǻ~3B`cR CȚOܡ(:⴫Zihz"* yޠIg˶N|`imȲ(y:/hRJidK |}K>J3֜e3JaXePւƸp" @Y-CN ЙXY`JƔccu!W?FbNZ3-zû. zHP.CV%T.1{C}6Ƃˤ|`o:dN <8(2 BeJVzݵ뮕^[iEq$DLKmPJ#ȡsRt[;UD*hQYc?4 5F6i6ݎJS54x2F'֭xMUE j6X5mep1kg@ " dI{*3 -9E$Fem{J-)V ,.D!Ԡ`+ax`˲=;y J5$%ZVTz=e=W5sW+Fޞƅ .2 mkVTiprP{ޑi$iBdoFr&57'׻7wY ]< ͠3tcݛKzi}t ݛ+>y"jkc Nw:՝W؄pi벻?hoz,eu|}F j+CzepE]۰P{JGtGD*$'!>qlR@&+d+AV" b [i[H_)cVeTЅ"R~an ֨- / _5vDS#E\iIl639 6h;Hg3T쇰3/''yejUY @buYmevaZl0eNS97:;V*6].XysBC}v϶4!-uBȹhYRH;9WF=Tȹ2dL~3OLCΥZ?I_X9R!dzPdRZ(j>JVX!fe[M+߆[e9x} =d3@Y~}BG)v_plEvBMBI;{^ꐠ`E 5 FT06?%HP2D&%dYSudf6"%HUyȿWKad #^hNkL]*&1^F/{ѭp ȹh&0h}Y!J֖wY %cdq-XI9cp~΄w)2/ogSNLnj`*"+iְ6 X⪒ܘFڃWj~I$cItteuyڴ`VUkFHۊ-(K yj{u%ȂN'[,Om+JFkH4^sȆDk$IUTT_ >;jM=]ŘT((BT5`J$ pRlADR>& %f BM K*IjrQ֙P"%8 xY!g3SiLvQ`4K4$o*lFa MJ0rл;;@wMj즸8KM1o8_y@çrr@Ayٞ!GZfJTbB/jqRi3R2n'CA]B*!Tyq\CILK։ҿ %B۫ ؽpSB8v=Rļ< wYeثҿCa!ԫtSy]"  s^h L庬L6(zejصԠu EfF2|vE~JˈiQ4$l!*Vؐhî@g$Wrrbݴ0#ɒh᫱6+l=zE;sO%kGƲZUs5TBL24KZ)iH{!rVhH$А`Vв\dtYUAFC=;}V( 1bm/խhHc}IR 9[~@CعCUJ`?xcaJJeQYiHGÔ"kYiH܁&!cG܄"}1dRwv5{ [+Ь^T"%i*v6Jb} ݶ)ѴtY4QkZ|@a?7ڳNCJRQK|44OCʧ4Y|4x4 itvc"R b(̋&bu61V'I'k^h"N-QI:):I(IQt|@KrrR4GqY-` .jk 2QP-EJ_&ˠ.hf !N@S9evqDW5o6w]F:ȟo]"I*PÕn>|on6O[?\?W~߼?ۼ7[R|n}0鿬oݰ]ru эkڿ! w!7u>֡A kG V-Q8I f' %x![5>꟭koԋPOU7vk;NI?&hSmPR&#?|hmiƷ%TvRXSbZE4(I3h$zcaMET˺Ȟ7ë`^}Xh&LU&&Ӄ"k bX,BRSUNZ[oǿ@/1oŋd8Y$هKR,Ki`{Ȟsnf򰱭V7Y^dWrL lhDHיj5R>6.c )r<$]<;*`VL9wҎKxl:o=$| \PZʺv\!U $Z="DR0@<%0@Y?KQǃNN{RPU7<Ԣ(t[A }*\+i-xQSfj`T'm?(`-+gR$>"=:p [P([ȝ,KJ,_%ȇ)3rcmxOL % Lao9epa. #yj+ᩅRbVϱtZplŪ* n`& YX൙vxlj )BRDhc0=MՃs AJ zp6W01zpF,a!=)X;Z/>AvKywc6X3Nf:92=W"ѠV7?M$|U%0iF̍r],}0uhhF>C  I:\m8Tҥ]4PHB~h:ЁD{<&%c,)?1y҄`ݓ^1>vWFy#.qyVoO\-]蘿EyF<΃44/cӼM26flB#"(a(oqh U"GSV0JBLN"r;f,&hVr\66`SӖNֵ7,8LF9GVt«$yG%@C&KUIb#SYh혿G&TJGxX`y窐t)zBʒҥQIP)Ny`55}[8B3gg:^G2|v9vz9D^<:-U !A )9Љ/%@:{-QBF:NJC3Հ3D Q;]Ɉ~cUhDRpr i+wEm^s-wEaw% P8ΐ{2V@]~L+ŌۿS}n'}3aؓb}jaY>@h#J. hpD'CdL!b(`[S+/08̛qÎ;^JR1[|MRP!p?'kXD 8>Qi7N"0:c6ݨzq7ⲉ̃UNMq^M%Ds*~ޯ_tvX/={h2t]Ga 1]8gĂC&Pp% G(T6˜Y@qs56[^XEH-m۹OQ І^bdGv-2˒mv7:V(#hr)ȳH2Q !Mv,be-%*@S #=8V2OHJoW쭤Jfr(i{>u8rQBh88h{1|zp5w'8hQ~I/Ny:)9⚥a3!)vkl<oؘpy.(1{r5U =;q O :@Z@q-2mmfHRIϬĴ)kĔ7a%T^)!+YF'ٷd~컔Z6:pGś;>[,l?W f$D٬gzfjW`UЎvaoZW# YC *0fCss )'rАt c",EaLct\%R3>5.NUYWrk5R>:U^˃Liuh b@yL#H}Mŵj%8.Ԑ<]Jߒ&K8pNٵ<,gE!]ёv5@@y̿R Q?GQyګ3x PLˡܝ*F?zDPg?_윧6{I۫χ"5۹^\h&kwOZP-q,cO|u 魻&yp/?ZS७DĂJidR>ھLXxQ <| rXβ 4):H+E%a\ mTWfehP#S3otr,)hczDLJ0@:k%;(  :XO)EIBHKm6lKR@iO+EAsA|PSwTrID Ձ.|GS}s1E Mb[w-] )?7W ^YwaO Db SK]~^ 0:z2(c}`%[MV&WXpC#_?ٹΨ @3Wd#lot2e:haXDMEZbomK+eJa"*EI(ӶTvCi>-,m}©$=<<-,KE6/SB_Z"̇cǓn˧Wqw1be(*0P6(|YONۗ@DEju*7\蛠U Df[;_^N (IN;A/]66&_&*8jwP[ZlqvPp68G.xi(ͨ@1hFFH9 s /<,2b-HŅ_&2Nou/a|8}եrG̓/vzrQc}6ɇ$|}/IY7 ca۴ ~z͛Ή;E3>&`j8V!Y{;qJ 唣>Y6@B ǨǶ8 Y2tEPQ-DbtI_cC "lxJҾ ے( H'mp(a<(ThuK1K taWU%(4A/XeE6!lx@k=ȒeZNfCS5[o/<IJ${7 69D ܓ7$={~bu>-w>龂{r{7Ӄ}+*HPXۘ1j*n`)/EIdTcJ,*dEDKU6歔OŔt\>ŴhB|Py4K(8E<9Hə).E6YRaTxr$ZK'hgxnI_&)N8Iw v`C^5$=u0HJqO 5gTBBR&H!,)ySdJkjli#PvWEI?ˉ{)GΨp⾕2} [}(-4zPZ5͐,LN7 ~?z8{p/zڼY}M MfǂFxrft~!L Xξ۹WNK| 9z6 4=,W^5MQu FI-T"èhC2Ng黔Z6Rvo7&AxxdvVyЮ T!@)bx!2IB ^9m9A#ʘ6]LBDluZ!؊TzK VmD(Vp5;Xŏ%mj8#7wjG4!!O+,$컶*}*q?ͣɻ=Oja[B)oY/n!)W]x_)Db懠i*wZ!a8~v-ze&Ia>fڷCCeZȉ7NmJgM Y)M7rF)FMh!'D)xskt0G6Hi:]3Llt\ѭrMuhJOXd|LLՉ'uX1 Ê3CUrL(HZ!ħyG_;y±ƋcjgO F;㧟Ep/%M>L 2]MӃAb39lZ^䷻& 96~}swO78Wۏzy^yQZ.7ζkj1~l\juS;&A.^E t9€.3+x7MRg[NZi*()S*MiZ4ߑBF[n߭ODH|mǶ/U۩oU&GJ(XAQyidYgc3_Ij>9U{UoLP]0Y$[ л>`\7*_MiZ^Wc;nF4JbEʌK,1X4l|bgFZ1L#B6yhlغh6( ejȑ1]N_)!;9&bLjeU']6IET^OXY,f,sz/ZWݳcKtLǍ`ɨJj/͠ahjp:R83yXdžayݱW6*Pr#z$z#9=6h8[+f;=VCsۑa9q|dV $ Q|~3&Ty1f+LἪQ[ d(9Hr ҵ(b&QHP&*a| LCE}"$5c = FaOؠ"60 #(,Ipria{lo% շTVVDK=ʢgD u>xv+}gLVHin:hGt\KuIS{v ^ȑZN>7SF4dجSMVRSl푼UDc ]Ԩw+j*Ո+0Sг+Zm!'h ;Vz)'Wk孩֦V*ɰ(-1ef$jZP&F)Ti*\fօf}PKؓ p{'Ţ8*,Dd"`V,4P'C 륋D8*K6D q<[--'k6K֝$^~}?}J?]@rsBq?' M?a՟;]& Sz(6x?!4oJ\onmԻvM@Ծ l+}ޛX(TxFMeT+3vOnz_jl̘C4˭yrIͼ 4 L򅑊H=aUu#4h˺&S/rm;7C ;(Vwt F9ZJCoH!kdLqJr.XX6w󊡹c@W*Wɢh ɪtX*"!͗E԰;/0*>ʛ ,r$t~w;YTyǷEA#Oct-"NwOΤvfYn*Q{ Վ E5WR- 5z^bUYYPQL:G󉠧V{/u* kV Uf9vZ5kVW~N%{]?!ĕ V/ +5+.zCRv_[ƦqRwm`Z;n7Z!{TMP܊f"uQw|sc:y7’/0*sAJ!0)t J)n]Qq77wc+ĉrgPûP0wKl[&eCl46+ l;u{>O*qhG+Oč"fz?^f8BBIE&.ǿ{؏I؜<ܥIoFV@^vש~.6cs隱ZLÏKз1ui_U7!u }'U*VL?J7i}AZfߺuw]la&}we^_AR[f;/~l{/GzF|WR 0VCSܱ( TrVϭ/C2c&IehSQaL"HHe10K\%1nӇFGϕ \DRoT,K)o5ȼދhtBB;I_F"Ν¤L 2ΦV{RDLf_rYNrY8xN.^!WH L5vxEGtK!Tk'*:{uG&Q-ApGAY Vؤ g~(We U{JrTҤ<&?9ձYLaqG4CNAts|2 Y!dXIم|p[e|s[7qt 񂦠\a('#4[ݨXV 繓+*ؑHWʠhI)J0*O}YhԡGM_alva X?#戣gO .AV "(E/Sah5 gI+QSݎWѴj jЫ_vS.CPau9XI^wJ^u9r9nݡUkX1SVͻ~2#.H#GK9hךM^b5k(MFiU_cW6/ 0$&a.~2U; ^uNsW}'6:ÅG*Dla`V֠^uRx4O |l.]ZԧqVm}~j̑|$ӇaW5dt:D@eIIs[c-`-vzK#)jՖQ&kߕt"\1ZB ҷDZ $"V)6jo"ڶuKZHƲ@'r*ڙVhC̘Vkz"M 2 к1|0eomcӰ Ej!QGi0+xcEZ/M= lFKRV(aVKpBZB䒰 6 ELHMkHtXʪV^%"F xLz'xLy"NEkLDu%K7tuR3#.8&gFa#&Viط2Rĸff8p!bT2d̡&7`r)y{MWD]t V\әH{sȸ h ?.]a/6Kr'L͊(MǺ;YKLNKKvZ^[h2 U 6i!,\#i*I3}cVf qM?M{a ΠS\',H34Hֺ~{v1Uț Kgz dDDU<ȞSdf&.Ѐ.[| %I@I ZFt Эӂb&WΚuS l4WM.3] 'ɶ,hR ƞaZ˚0yj=+FB@Y ZmB,J/.J=PR80"( D\%5Oֽj'4ml ٦;XxS2>pIxhis?\DŽom 7k 6^AlԮ7ڷ]ZuӎTCvL)\C($ʆ]*du*U2їo.CwM)p]/ J_[4ޮݘ j1[6ۛɝN6-Y%8[w>쭠c1.h"sɫ7\JȚC/^B{egAbKJ)XБw\R̦vf$D]X!i9,ƙ2a=L`D}ˢMZ)WRزm$Lrm kwra6zC]Is\7+ _Ё"\KwxgXeJdb,ڿ~EJ*XLT-0/\L 5PTZ?n%ݺME"CE$wZ[cC3v&*oDNuM|w8uV bynX\)*X?&fy<-sA)*-X?4sF3۞ ֛̻A6z51|>"g(WSx&qEyW܎VΚ^5(O?~|;YK/Cpot̜ fӵW{Տ/k}뻾3tT#_~u<"C^꿮~y85q!#7 6 Ý]5`5Т a<%iT."UizMcnfFav}c΅lNQ*+]J|KɆw(JK-pزB]mA匮8U 7] `N .SK.1xJ 5 viT *q\zZg[leZ4!jK{ CvI5 ;T3.t2Ceu!n=מGi{Uemz^TkՉ-^sYv>iXs.G!(ʲԨaΖ+0瞝BNšg棱.u1RMBsśk4@9  VGCۺPKZd'0:71%+QϛUJK !miCj-8W5d肌m97U!fIP{6rhDvKisw|*:J?Wb9vMu'1WǏֱj}z{9alQߧ;8g֢͊̃NΪWl׷QWS5 3HMf ñvlz4#;SAn@iUGi8dg4\g'6s!X ih#[㟽y@˛Lg{h. ;סzPum^ɡ*dR`4Fs[O{A 4 0  1(bddO{}Ĵl<£o>i(0:81³ : ^{ $J؋CQ|է$`K S.V:uWRm+c u |11t&to="N!μA Ή'M×Ch v<{l4igNgi.qIO/EmffqYPS'P508Nb袽>^.~[k9 ֱ/,Lξ^&{utҫx 2Bo.>DPf@ QTpY{M2oK/4 _/ ! ?{ˢ5i9Ngwx0ƥ9[ (]cP$S^wjr-_\ΡSװ!L &ʦ 0r B=i8FC3 V<{!·(@֓<{B ^lp}B}@>ѯuIuT\1X)SZ.p5 5nd5`F3mh>{1]{W.n<5𮚔;3TPպb' 9n3طlWv,8C#{WX%S;EtecK`Z:`ٗ%DįJ/˹[4%2๘T]6CjAo7y;TΚD,j-@ʩ5Mu3ڑ껅4 tj诔yVch ZMQSko[٠p[);SN/W$GܺJ{Ȇϗ^uihzr껺9 6p;T-qG;Q_iռZuUK; 7D|ٷoܥ>;6AEQLd¯r]r 1ߎ֫|*6o_0ɹCQ˯TAY-}jި7GYJf?=;™NwWQ]!㤙JBt37JrIW>lp,Y#tx$i8i $l&R7->iޕ1 &CӅAH~Q Fk?SSQIk誺`NŜ_]3r?:5> w u|O;)9eOUwIQz:3eyrtݪFQyqRnVyB 5F\h"Apkѧ<}7_{>PXr/Qb6J_q^tj_WSm=rz42 i\ _G&3g[bKg;>IuQ6mʯlڡWtͬ+U|V:|v hw: m)7Q`7coec/zoYmʜR͹X5/WD kv4ji\Plk" q XXZ.V.W?݅v4ҏǛhg:o#H",Ǹqnvӡ8 I~rg@tw ҇ޭ)ƍޞ}>}.win;SR_( a byH?q|,o׼.w؝D(8s^OǟӮEF~?& Í?ߡFo7{lɒXMƦK >sepg{YXEs|nFn[R) #Uϩw¶.B=RFҥ%l;HM klk%a=c,@ERo[h{@ziԔ=l/53nWd\DVt5^ŪImp>}eP^|9lY7)yyTkĕbj֖]p9HE􅧀Ԍt'ZVfáZ]dT( {SAlND؍E-\`wz J=$y8sF} ; д\ xFZ 4R^hDD'5gFɬ#bB',k?@JΤt^H`$(S'f,^]:'bqLyZ]H{cԨ4M`COaZ/g7i0[i5x3,NN0#F3kОMnzfln2XY6SRD" ٴMfy &C..6 [i|ke沩wZ` rZPdRLNe;{Z'$d!RTi»#Ӑqx9:S4JM$F+k5WzVWUuKUks)tqtCԏGi f-j hgx\Sˆ*IF,x[Y]b.`I%HUXoG9gE:/W[^:AFa<+H<"Tw"WhQ/A,"Ir`+Dkz.qvV'Lԕ"VM*#EI~FI[72J I#ɥ3bt#2u~0ψ.F"S>pFe[^)@>ҰS.e@޿ ?Oʁ+*|i["8탅(ɾ;FodG_]Sq*|2NPj fFS̑ 4 g X VZ> ق1?JE>]<5>E-Њ%o ]ȌtФWQ#!("iIw7}862&(ZcBڜ Zq9DSg$ (u rVVzsI`,Sf8-pMI (ȁp0(OT胋DtN|,5ڔˡp@_RيApLzG&jM(4}fQ VֆS1c`2`B|=qE ~G?mlٯn>ȋ`Ru҇'O8thE4*&n0na~{#s`2n\Ȩ٪=w-RCd^@/]gᏴx%;M@bl[L ) )?]&Wdgy wۏlY^F(mex3|t: lwbJw9˜O8:[UyRnfev)K)4 ~cR jqn:{k$$0_t=urr)lD6m=8J¯yCԢԱ!W&:$[D Y%sʩ` ^*MFft\(9B@!-|#(@tՈWlB-e &9%bD4,SvGCOsݨCY km*]݀jz24RϬ[+H.~9i8& 8ހ]o #XF GHnb~?ƹ&W^)l_;74[~7/u:n"c4x5Yn?n+2 b6b)Z"Wn>zӂ90[8G.1e&,4~#lu8|鿖!zeX[tmIzN!I!xPo e򗗭S V ?K-!"B+(~+uCYa\tkDpE$))z>-o7?]81J\F/u1jdޚ(Z?'Gode#-eA'^L.\d]Y0f"-G+ yI0gjv0GR-Y%(,0TYiIRy $ctp嬲Pn˄kQm2mZX6(bde(@d.qfTTk͑g%GsiV8IAJd)ųEH!i@}*#ZXLj:9f#l @z u!3 IQ!y+&uw9x ':B#}^Yp%͍(%ZS˒2X؂j\ko4tvZ3kooOnנkέO}Ae;ԿHXq&!2=D'1 di4mֈ] sL1W<ݦNul 6GemPߝ&'O/~)OsKꉒJPF1&Rtԧ̬2B'gwv#j_M騅pi{<6߉i$KIvxT ^[?IZF7JUCF;p~,ޯˡ_-0;>ךbX>Khc6-ŧ) Ӥʃ阰T u'е9G8%M}菝Hm5.R[==w1RZgb5dָ6gq oR-Amu ذ_ZO@Xw7%tzw׋_"=t2 .yP:Kl\,?h}#t2xNwˏ($[UE;2 ˱2Gp۴-KA>J{=(Q.FҐW-:z?UOmݴ`ZyP:cݎ}4nnhm y*ڤSC>&8bukA괾u;.B13ukOznch+W Drhgj`dpjFN+3 dgAk2cB2D0&3CpFN #6Q Ogґ!:M_J%n Ɏcf9m lDOwGl- כ',6NW8f^_K+[L5!]_%)ezd?3%5~6lLQT^4 >W-a)|4UsߢvT6(=5Q.sZxo &(utlf 7DE$çC$ƞ*yvkJdIQ{^'rZd2d)mKzkfLg҇u6z+y*"f:DB8M`.j/CX@*1wzTJV;Gz}SM]xf@_XL +֏$ ssoDܐW 1\q Gސ KF ]9ZqѥDKܯ:\=cF@v{ $gl~B+hhυX[GaZo:.a6ܬ JɄUZlZ)d-lR̹! nKB*e7tAZb²D_SCuդ=)Z)?@ta^&nB֑gq$5JGdYT: ^}bXwP:suT;-H{v`)Ii *y : -~wF=x8sVX?k8aW97=ګ'l&:IՐ``Ru.b &Up-$؊ zITk&sǣr #aebŴkRۓ9mzΙ5lVn=WTvU1yr}&H;5y@. C&D˫ar>w½hÄ}(Ɵ-H8Y' T^tK1Wȥԓ5rZڽ[8\S;M g8d8)Ն1A&<"p+jV[x6ȴcqh x y*ڤSSMFVԨN;X^$Snm y*ڢS}F[֭$74 7mA26uk-0rЩʼMsA]w\Nt;s[Vb_sC[$k"$DJAoVsƓt1˼w)+|2)ݢL6^TGD%mP;5z]9TeDJIY"BEC^jcoCf,c$x RL,e.ȈtY3Q T "O l'9Yg- y&Z$@ƒz`6(tFGP>j! : msіȍGcD +iӎG )b5PTa6҃5lE&`=D iNR0YD3'ɂd]sé- 1GQ+ "@zQ@;ф)ޥb+g" T[v`IYK[2oDΊ%pd"O:L5jj1*12vbBDCwr.#LB言L-aT ewSG*ߺ7~q= w3z-MvVzM08]ԾߤVdcHrͽ*{ӟWd+0Eb;X"y=EotMM'S!nVsmPG1R)!NLnA[J#Q&jGHlS_~J6-e[W<ɊޞH w1ew~"6ȄJ2a^1;NTiCҩA(Ϩ1m0S^)9hOj(J˞N{6_iGTQi#PōgyipmTΎIn'1JcFUFg LAޞ}YcRV4sSd%\!S8 0]?ZZ,Tp 5NW?Uuq]]pmQ{I➵O/MZU2G͹qF\0Q0_NDM}̪;*?ǐn35uvMx~*ZTz/g\+H$X .nLTBĄy j&]3u!D'x>f)X"TZ$AX'4t>3t%EeNfZ!ME2 hp8ը7ӱ"j"1ci!ĬBA迒BFNeȘiGU!Q #&!,1rZ9 V[R'y6DxIgԮfW}ήhpwS O\6K!.(@- ߦH ],n޾yIsxKoVʥ<$^˳E9{v'BrazWҖB&+OƏ=T|--a@+ɲDJBT/&CYTr2p+akbJwUI| <\*M+|/&+9#(] qvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003434034615153516164017713 0ustar rootrootMar 09 09:19:57 crc systemd[1]: Starting Kubernetes Kubelet... Mar 09 09:19:58 crc restorecon[4691]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 09 09:19:58 crc restorecon[4691]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 09 09:19:59 crc kubenswrapper[4692]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 09:19:59 crc kubenswrapper[4692]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 09 09:19:59 crc kubenswrapper[4692]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 09:19:59 crc kubenswrapper[4692]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 09:19:59 crc kubenswrapper[4692]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 09 09:19:59 crc kubenswrapper[4692]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.765437 4692 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768917 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768938 4692 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768945 4692 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768950 4692 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768956 4692 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768961 4692 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768966 4692 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768971 4692 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768976 4692 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768981 4692 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768988 4692 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.768994 4692 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769000 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769005 4692 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769010 4692 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769015 4692 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769021 4692 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769028 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769033 4692 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769040 4692 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769047 4692 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769053 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769060 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769066 4692 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769071 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769076 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769082 4692 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769087 4692 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769098 4692 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769103 4692 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769108 4692 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769113 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769117 4692 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769122 4692 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769127 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769132 4692 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769136 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769141 4692 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769146 4692 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769151 4692 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769157 4692 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769185 4692 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769192 4692 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769197 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769203 4692 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769215 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769223 4692 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769230 4692 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769236 4692 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769242 4692 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769248 4692 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769254 4692 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769260 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769266 4692 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769271 4692 feature_gate.go:330] unrecognized feature gate: Example Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769276 4692 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769283 4692 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769288 4692 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769292 4692 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769297 4692 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769302 4692 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769307 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769312 4692 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769316 4692 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769321 4692 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769327 4692 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769334 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769339 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769345 4692 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769350 4692 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.769357 4692 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.769992 4692 flags.go:64] FLAG: --address="0.0.0.0" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770010 4692 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770023 4692 flags.go:64] FLAG: --anonymous-auth="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770030 4692 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770037 4692 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770043 4692 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770051 4692 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770058 4692 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770064 4692 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770069 4692 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770076 4692 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770083 4692 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770088 4692 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770094 4692 flags.go:64] FLAG: --cgroup-root="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770099 4692 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770105 4692 flags.go:64] FLAG: --client-ca-file="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770111 4692 flags.go:64] FLAG: --cloud-config="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770117 4692 flags.go:64] FLAG: --cloud-provider="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770122 4692 flags.go:64] FLAG: --cluster-dns="[]" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770129 4692 flags.go:64] FLAG: --cluster-domain="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770134 4692 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770140 4692 flags.go:64] FLAG: --config-dir="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770145 4692 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770151 4692 flags.go:64] FLAG: --container-log-max-files="5" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770158 4692 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770199 4692 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770205 4692 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770211 4692 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770217 4692 flags.go:64] FLAG: --contention-profiling="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770223 4692 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770228 4692 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770234 4692 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770241 4692 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770248 4692 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770254 4692 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770260 4692 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770265 4692 flags.go:64] FLAG: --enable-load-reader="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770271 4692 flags.go:64] FLAG: --enable-server="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770277 4692 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770285 4692 flags.go:64] FLAG: --event-burst="100" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770290 4692 flags.go:64] FLAG: --event-qps="50" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770296 4692 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770301 4692 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770307 4692 flags.go:64] FLAG: --eviction-hard="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770314 4692 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770320 4692 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770326 4692 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770332 4692 flags.go:64] FLAG: --eviction-soft="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770338 4692 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770344 4692 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770351 4692 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770358 4692 flags.go:64] FLAG: --experimental-mounter-path="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770365 4692 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770373 4692 flags.go:64] FLAG: --fail-swap-on="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770380 4692 flags.go:64] FLAG: --feature-gates="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770389 4692 flags.go:64] FLAG: --file-check-frequency="20s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770395 4692 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770402 4692 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770409 4692 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770416 4692 flags.go:64] FLAG: --healthz-port="10248" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770423 4692 flags.go:64] FLAG: --help="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770430 4692 flags.go:64] FLAG: --hostname-override="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770437 4692 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770444 4692 flags.go:64] FLAG: --http-check-frequency="20s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770451 4692 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770458 4692 flags.go:64] FLAG: --image-credential-provider-config="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770464 4692 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770471 4692 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770478 4692 flags.go:64] FLAG: --image-service-endpoint="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770484 4692 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770491 4692 flags.go:64] FLAG: --kube-api-burst="100" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770498 4692 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770506 4692 flags.go:64] FLAG: --kube-api-qps="50" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770512 4692 flags.go:64] FLAG: --kube-reserved="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770520 4692 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770526 4692 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770533 4692 flags.go:64] FLAG: --kubelet-cgroups="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770540 4692 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770547 4692 flags.go:64] FLAG: --lock-file="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770554 4692 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770563 4692 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770570 4692 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770589 4692 flags.go:64] FLAG: --log-json-split-stream="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770596 4692 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770603 4692 flags.go:64] FLAG: --log-text-split-stream="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770610 4692 flags.go:64] FLAG: --logging-format="text" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770616 4692 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770623 4692 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770628 4692 flags.go:64] FLAG: --manifest-url="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770634 4692 flags.go:64] FLAG: --manifest-url-header="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770642 4692 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770648 4692 flags.go:64] FLAG: --max-open-files="1000000" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770655 4692 flags.go:64] FLAG: --max-pods="110" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770660 4692 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770666 4692 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770672 4692 flags.go:64] FLAG: --memory-manager-policy="None" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770677 4692 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770683 4692 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770688 4692 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770694 4692 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770706 4692 flags.go:64] FLAG: --node-status-max-images="50" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770712 4692 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770718 4692 flags.go:64] FLAG: --oom-score-adj="-999" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770724 4692 flags.go:64] FLAG: --pod-cidr="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770729 4692 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770739 4692 flags.go:64] FLAG: --pod-manifest-path="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770744 4692 flags.go:64] FLAG: --pod-max-pids="-1" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770750 4692 flags.go:64] FLAG: --pods-per-core="0" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770756 4692 flags.go:64] FLAG: --port="10250" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770761 4692 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770767 4692 flags.go:64] FLAG: --provider-id="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770773 4692 flags.go:64] FLAG: --qos-reserved="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770778 4692 flags.go:64] FLAG: --read-only-port="10255" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770784 4692 flags.go:64] FLAG: --register-node="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770790 4692 flags.go:64] FLAG: --register-schedulable="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770795 4692 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770804 4692 flags.go:64] FLAG: --registry-burst="10" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770810 4692 flags.go:64] FLAG: --registry-qps="5" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770816 4692 flags.go:64] FLAG: --reserved-cpus="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770821 4692 flags.go:64] FLAG: --reserved-memory="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770828 4692 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770834 4692 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770840 4692 flags.go:64] FLAG: --rotate-certificates="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770845 4692 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770851 4692 flags.go:64] FLAG: --runonce="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770857 4692 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770863 4692 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770868 4692 flags.go:64] FLAG: --seccomp-default="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770874 4692 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770879 4692 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770885 4692 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770890 4692 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770899 4692 flags.go:64] FLAG: --storage-driver-password="root" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770905 4692 flags.go:64] FLAG: --storage-driver-secure="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770910 4692 flags.go:64] FLAG: --storage-driver-table="stats" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770916 4692 flags.go:64] FLAG: --storage-driver-user="root" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770921 4692 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770927 4692 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770933 4692 flags.go:64] FLAG: --system-cgroups="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770938 4692 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770946 4692 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770952 4692 flags.go:64] FLAG: --tls-cert-file="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770958 4692 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770964 4692 flags.go:64] FLAG: --tls-min-version="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770970 4692 flags.go:64] FLAG: --tls-private-key-file="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770975 4692 flags.go:64] FLAG: --topology-manager-policy="none" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770981 4692 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770987 4692 flags.go:64] FLAG: --topology-manager-scope="container" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770992 4692 flags.go:64] FLAG: --v="2" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.770999 4692 flags.go:64] FLAG: --version="false" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.771007 4692 flags.go:64] FLAG: --vmodule="" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.771013 4692 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.771019 4692 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772000 4692 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772014 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772020 4692 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772026 4692 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772032 4692 feature_gate.go:330] unrecognized feature gate: Example Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772037 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772042 4692 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772046 4692 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772053 4692 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772059 4692 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772065 4692 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772073 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772078 4692 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772083 4692 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772088 4692 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772093 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772098 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772103 4692 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772108 4692 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772112 4692 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772117 4692 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772123 4692 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772128 4692 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772133 4692 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772138 4692 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772143 4692 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772148 4692 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772153 4692 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772158 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772186 4692 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772192 4692 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772196 4692 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772201 4692 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772206 4692 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772210 4692 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772215 4692 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772222 4692 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772229 4692 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772236 4692 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772242 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772247 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772252 4692 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772257 4692 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772264 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772269 4692 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772274 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772278 4692 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772283 4692 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772288 4692 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772293 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772299 4692 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772305 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772311 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772315 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772320 4692 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772325 4692 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772330 4692 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772335 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772340 4692 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772345 4692 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772349 4692 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772354 4692 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772359 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772364 4692 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772369 4692 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772373 4692 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772378 4692 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772383 4692 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772388 4692 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772393 4692 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.772397 4692 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.772413 4692 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.785975 4692 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.786022 4692 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786249 4692 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786272 4692 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786285 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786298 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786309 4692 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786322 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786334 4692 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786349 4692 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786369 4692 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786385 4692 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786400 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786413 4692 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786426 4692 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786439 4692 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786451 4692 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786462 4692 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786477 4692 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786489 4692 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786501 4692 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786511 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786522 4692 feature_gate.go:330] unrecognized feature gate: Example Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786533 4692 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786544 4692 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786555 4692 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786565 4692 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786577 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786588 4692 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786600 4692 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786612 4692 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786623 4692 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786634 4692 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786645 4692 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786656 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786667 4692 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786681 4692 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786692 4692 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786703 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786714 4692 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786725 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786736 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786747 4692 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786758 4692 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786769 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786780 4692 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786792 4692 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786802 4692 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786813 4692 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786824 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786835 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786846 4692 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786857 4692 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786868 4692 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786879 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786894 4692 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786909 4692 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786923 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786936 4692 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786949 4692 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786961 4692 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786975 4692 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.786989 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787000 4692 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787013 4692 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787025 4692 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787036 4692 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787049 4692 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787059 4692 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787070 4692 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787081 4692 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787092 4692 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787106 4692 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.787124 4692 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787531 4692 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787553 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787569 4692 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787586 4692 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787598 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787610 4692 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787622 4692 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787634 4692 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787646 4692 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787657 4692 feature_gate.go:330] unrecognized feature gate: Example Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787668 4692 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787679 4692 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787689 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787701 4692 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787711 4692 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787722 4692 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787733 4692 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787744 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787755 4692 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787765 4692 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787778 4692 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787789 4692 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787800 4692 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787810 4692 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787822 4692 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787833 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787844 4692 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787855 4692 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787866 4692 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787877 4692 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787888 4692 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787899 4692 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787909 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787920 4692 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787946 4692 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787959 4692 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787969 4692 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787979 4692 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.787989 4692 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788000 4692 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788010 4692 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788020 4692 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788031 4692 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788041 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788052 4692 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788062 4692 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788073 4692 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788083 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788093 4692 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788105 4692 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788116 4692 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788126 4692 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788141 4692 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788151 4692 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788193 4692 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788205 4692 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788216 4692 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788227 4692 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788238 4692 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788248 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788259 4692 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788274 4692 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788288 4692 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788300 4692 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788315 4692 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788328 4692 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788340 4692 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788350 4692 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788361 4692 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788371 4692 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.788384 4692 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.788401 4692 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.788738 4692 server.go:940] "Client rotation is on, will bootstrap in background" Mar 09 09:19:59 crc kubenswrapper[4692]: E0309 09:19:59.799905 4692 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.804687 4692 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.804835 4692 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.809189 4692 server.go:997] "Starting client certificate rotation" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.809277 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.809429 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.849872 4692 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.853954 4692 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 09 09:19:59 crc kubenswrapper[4692]: E0309 09:19:59.857695 4692 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.886724 4692 log.go:25] "Validated CRI v1 runtime API" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.950662 4692 log.go:25] "Validated CRI v1 image API" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.952694 4692 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.959369 4692 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-09-09-16-06-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.959409 4692 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.980862 4692 manager.go:217] Machine: {Timestamp:2026-03-09 09:19:59.978044176 +0000 UTC m=+0.802779777 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:152c1cda-659c-435d-a04c-ef7c06479b86 BootID:09ec62b1-ccf9-4248-acae-6d6e8481578e Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ec:a0:ce Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ec:a0:ce Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e6:ed:e8 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:dc:99:61 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c3:61:70 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c8:9d:f0 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:52:b8:f7:83:04:15 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7a:99:69:89:22:9e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.981186 4692 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.981418 4692 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.983229 4692 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.983481 4692 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.983521 4692 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.983763 4692 topology_manager.go:138] "Creating topology manager with none policy" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.983775 4692 container_manager_linux.go:303] "Creating device plugin manager" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.984280 4692 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.984317 4692 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.984614 4692 state_mem.go:36] "Initialized new in-memory state store" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.984726 4692 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.988779 4692 kubelet.go:418] "Attempting to sync node with API server" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.988809 4692 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.988838 4692 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.988857 4692 kubelet.go:324] "Adding apiserver pod source" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.988874 4692 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.994500 4692 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 09 09:19:59 crc kubenswrapper[4692]: I0309 09:19:59.996110 4692 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.997121 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:19:59 crc kubenswrapper[4692]: E0309 09:19:59.997260 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:19:59 crc kubenswrapper[4692]: W0309 09:19:59.997288 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:19:59 crc kubenswrapper[4692]: E0309 09:19:59.997362 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.000233 4692 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002432 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002466 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002473 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002482 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002495 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002504 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002513 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002525 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002535 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002543 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002554 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.002561 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.005536 4692 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.006117 4692 server.go:1280] "Started kubelet" Mar 09 09:20:00 crc systemd[1]: Started Kubernetes Kubelet. Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.007860 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.008147 4692 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.008136 4692 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.008865 4692 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.012257 4692 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.012310 4692 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.014657 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.015090 4692 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.015188 4692 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.015232 4692 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.015350 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="200ms" Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.015943 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.016012 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.016264 4692 factory.go:55] Registering systemd factory Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.016295 4692 factory.go:221] Registration of the systemd container factory successfully Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.016344 4692 server.go:460] "Adding debug handlers to kubelet server" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.023453 4692 factory.go:153] Registering CRI-O factory Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.023524 4692 factory.go:221] Registration of the crio container factory successfully Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.023653 4692 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.023700 4692 factory.go:103] Registering Raw factory Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.023738 4692 manager.go:1196] Started watching for new ooms in manager Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.027395 4692 manager.go:319] Starting recovery of all containers Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.026784 4692 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.181:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189b21c312144c93 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,LastTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036307 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036375 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036396 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036415 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036432 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036449 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036464 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036479 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036500 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036514 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036528 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036545 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036560 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036579 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036595 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036610 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036648 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036662 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036675 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036687 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036700 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036714 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036726 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036741 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036754 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036769 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036785 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036799 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036812 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036825 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036838 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036850 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036870 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036915 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036971 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.036987 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037000 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037012 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037024 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037036 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037049 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037087 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037100 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037112 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037124 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037135 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037148 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037177 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037192 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037207 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037220 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037233 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037249 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037263 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037276 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037289 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037303 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037318 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037329 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037342 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037354 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037366 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037379 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037391 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037404 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037415 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037427 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037440 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037453 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037466 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037479 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037492 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037504 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037540 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037553 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037565 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037576 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037589 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037602 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037614 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037628 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037641 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037654 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037667 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037679 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037694 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037707 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037722 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037734 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037746 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037757 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037769 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037787 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037800 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037812 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037823 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037837 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037850 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037862 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037874 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037885 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037898 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037910 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037921 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037939 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037951 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037963 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037976 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.037988 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038001 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038014 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038027 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038041 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038055 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038067 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038082 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038094 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038108 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038120 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038133 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038145 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038158 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038186 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038200 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038215 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038235 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038252 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038267 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038285 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038298 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038310 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038323 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038337 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038350 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038362 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038433 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038445 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038459 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038476 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038490 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038504 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038543 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038578 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038592 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038606 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038618 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038632 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038646 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038658 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038671 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038686 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038700 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038714 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038726 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038740 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038753 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038765 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038778 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038791 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038803 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038817 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038829 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038842 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038855 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038868 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038881 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038893 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038906 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038918 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.038930 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041067 4692 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041096 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041110 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041121 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041131 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041141 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041153 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041182 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041198 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041209 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041220 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041229 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041239 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041248 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041259 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041268 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041280 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041296 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041309 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041325 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041338 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041351 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041365 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041375 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041386 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041396 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041406 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041415 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041424 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041433 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041444 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041454 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041465 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041475 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041486 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041496 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041536 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041546 4692 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041556 4692 reconstruct.go:97] "Volume reconstruction finished" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.041562 4692 reconciler.go:26] "Reconciler: start to sync state" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.047288 4692 manager.go:324] Recovery completed Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.058312 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.059662 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.059707 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.059721 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.060657 4692 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.060678 4692 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.060698 4692 state_mem.go:36] "Initialized new in-memory state store" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.067750 4692 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.070226 4692 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.070278 4692 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.070315 4692 kubelet.go:2335] "Starting kubelet main sync loop" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.070461 4692 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.074000 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.074109 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.092128 4692 policy_none.go:49] "None policy: Start" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.093297 4692 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.093323 4692 state_mem.go:35] "Initializing new in-memory state store" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.115267 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.157455 4692 manager.go:334] "Starting Device Plugin manager" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.157506 4692 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.157518 4692 server.go:79] "Starting device plugin registration server" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.158390 4692 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.158414 4692 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.158719 4692 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.158968 4692 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.159072 4692 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.167690 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.171109 4692 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.171297 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.173082 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.173124 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.173135 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.173337 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.174996 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.175034 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176258 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176302 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176312 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176409 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176465 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176476 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176527 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176553 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.176466 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177424 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177459 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177468 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177633 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177659 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177668 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177824 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.177979 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.178018 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.178932 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.178991 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.179007 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.179271 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.179298 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.179307 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.179451 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.179550 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.179592 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.180866 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.180890 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.180898 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.180874 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.180999 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.181009 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.181186 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.181213 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.181811 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.181882 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.181893 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.216502 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="400ms" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243682 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243720 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243746 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243764 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243780 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243796 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243814 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243830 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243848 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243863 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243878 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243900 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243915 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243929 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.243944 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.259299 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.260692 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.260728 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.260738 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.260765 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.261238 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345232 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345329 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345377 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345423 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345460 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345492 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345493 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345559 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345598 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345527 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345515 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345598 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345569 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345699 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345666 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345729 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345744 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345774 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345776 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345819 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345835 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345852 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345875 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345897 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345917 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345960 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345952 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.346003 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.346093 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.345860 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.462018 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.464513 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.464569 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.464588 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.464625 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.465374 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.506239 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.511575 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.529584 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.552123 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.556647 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.570171 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-50fd54771a2d6150e5d9621031e7316e5211c2bf3f12bce0063e2c06b5ccc817 WatchSource:0}: Error finding container 50fd54771a2d6150e5d9621031e7316e5211c2bf3f12bce0063e2c06b5ccc817: Status 404 returned error can't find the container with id 50fd54771a2d6150e5d9621031e7316e5211c2bf3f12bce0063e2c06b5ccc817 Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.572073 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-32eedccf3d3ccaa9c3068f385b65282d7ba7af8d19674e2135d6b1dd1543921c WatchSource:0}: Error finding container 32eedccf3d3ccaa9c3068f385b65282d7ba7af8d19674e2135d6b1dd1543921c: Status 404 returned error can't find the container with id 32eedccf3d3ccaa9c3068f385b65282d7ba7af8d19674e2135d6b1dd1543921c Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.585791 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-e76225c44662981dde9dfe13f5c07117a766d8894ae492cfb004c316ebcbde01 WatchSource:0}: Error finding container e76225c44662981dde9dfe13f5c07117a766d8894ae492cfb004c316ebcbde01: Status 404 returned error can't find the container with id e76225c44662981dde9dfe13f5c07117a766d8894ae492cfb004c316ebcbde01 Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.586511 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-f9fe74169f9c4363a4d8aad5a6bb0f0e9b5012e74f4abf99e111c14aac1db7d2 WatchSource:0}: Error finding container f9fe74169f9c4363a4d8aad5a6bb0f0e9b5012e74f4abf99e111c14aac1db7d2: Status 404 returned error can't find the container with id f9fe74169f9c4363a4d8aad5a6bb0f0e9b5012e74f4abf99e111c14aac1db7d2 Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.597798 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-4d32a36ca89ba09f237fcdd8c31db10070b62cbd47657278e605e155ccaf95b1 WatchSource:0}: Error finding container 4d32a36ca89ba09f237fcdd8c31db10070b62cbd47657278e605e155ccaf95b1: Status 404 returned error can't find the container with id 4d32a36ca89ba09f237fcdd8c31db10070b62cbd47657278e605e155ccaf95b1 Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.618061 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="800ms" Mar 09 09:20:00 crc kubenswrapper[4692]: W0309 09:20:00.835570 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.835659 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.865513 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.866655 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.866689 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.866699 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:00 crc kubenswrapper[4692]: I0309 09:20:00.866720 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:00 crc kubenswrapper[4692]: E0309 09:20:00.867004 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.009491 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:01 crc kubenswrapper[4692]: W0309 09:20:01.074413 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:01 crc kubenswrapper[4692]: E0309 09:20:01.074517 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.075691 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e76225c44662981dde9dfe13f5c07117a766d8894ae492cfb004c316ebcbde01"} Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.077041 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f9fe74169f9c4363a4d8aad5a6bb0f0e9b5012e74f4abf99e111c14aac1db7d2"} Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.078808 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"32eedccf3d3ccaa9c3068f385b65282d7ba7af8d19674e2135d6b1dd1543921c"} Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.079870 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"50fd54771a2d6150e5d9621031e7316e5211c2bf3f12bce0063e2c06b5ccc817"} Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.080794 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4d32a36ca89ba09f237fcdd8c31db10070b62cbd47657278e605e155ccaf95b1"} Mar 09 09:20:01 crc kubenswrapper[4692]: W0309 09:20:01.305896 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:01 crc kubenswrapper[4692]: E0309 09:20:01.305974 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:01 crc kubenswrapper[4692]: E0309 09:20:01.418690 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="1.6s" Mar 09 09:20:01 crc kubenswrapper[4692]: W0309 09:20:01.505412 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:01 crc kubenswrapper[4692]: E0309 09:20:01.505504 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.667434 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.668953 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.668998 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.669015 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:01 crc kubenswrapper[4692]: I0309 09:20:01.669046 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:01 crc kubenswrapper[4692]: E0309 09:20:01.669653 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.008971 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.016208 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 09:20:02 crc kubenswrapper[4692]: E0309 09:20:02.016913 4692 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.085930 4692 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fda1ec14b76051ff482d211cf87902c27bb66b00b9884610a5d2e2b5d208d02b" exitCode=0 Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.086007 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fda1ec14b76051ff482d211cf87902c27bb66b00b9884610a5d2e2b5d208d02b"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.086067 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.089201 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.089237 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.089254 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.092455 4692 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="64ccaf933dae9cd1f75bae51bb99f89adad701dbf2ad86b70e7ab3dda7a6d55b" exitCode=0 Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.092569 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"64ccaf933dae9cd1f75bae51bb99f89adad701dbf2ad86b70e7ab3dda7a6d55b"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.093148 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.094299 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.094333 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.094345 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.096172 4692 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ea3bc77f7daa482a9f64f96f03dc0a817786a96df0b608d9fb1fb94621897264" exitCode=0 Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.096230 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ea3bc77f7daa482a9f64f96f03dc0a817786a96df0b608d9fb1fb94621897264"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.096316 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.097023 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.097050 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.097058 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.100856 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4ad10811dca21c5e5479068d14d19e443a4b8a5c374f02326b02dc935116a5d8"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.100942 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"de87b6945eb66f7ccd3e4f1ac81bb7aab7950eaaa87b3c1d55733f7daa7f92d2"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.100974 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c3b766af9a91908ad63a850d6e36a104fb38f31802759317db0db0a364140055"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.101000 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0c021cb088029d8da4b3a5fee28a92cd55b25ceb658789d94bf42b674993f985"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.101243 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.103007 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.103062 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.103086 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.103511 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769" exitCode=0 Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.103555 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.103574 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769"} Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.104538 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.104565 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.104574 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.108272 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.109239 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.109299 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:02 crc kubenswrapper[4692]: I0309 09:20:02.109324 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:02 crc kubenswrapper[4692]: W0309 09:20:02.797882 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:02 crc kubenswrapper[4692]: E0309 09:20:02.798029 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:02 crc kubenswrapper[4692]: W0309 09:20:02.832595 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:02 crc kubenswrapper[4692]: E0309 09:20:02.832711 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.008659 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:03 crc kubenswrapper[4692]: E0309 09:20:03.021804 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="3.2s" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.109535 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ea319a43c4caf0cdea8d3da58dda0a348dc74efb4900c66205d3e679737f4857"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.109612 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e145a337783c6b10511df76b005c784e55701fa0e50f702c5e04bb97a199f56a"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.109628 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c48ad3a6894fbfe4d1c9cad9314586833ae3ce48922f1c629d452771f5dc36c5"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.109705 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.111196 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.111252 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.111274 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.114073 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.114133 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.114152 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.114195 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.116407 4692 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e949fcf15a21c500cf4cd6099a9d88e93d245c20380a9da8602219f5eb7e0bd2" exitCode=0 Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.116529 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e949fcf15a21c500cf4cd6099a9d88e93d245c20380a9da8602219f5eb7e0bd2"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.116789 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.118354 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.118389 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.118439 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.119518 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.120094 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6ff80c5c1de1e1669fb2f0a0dacf62d2f7bc4c969f671e336e959da467d6efa6"} Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.120192 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.120901 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.120944 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.120955 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.123673 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.123731 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.123752 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:03 crc kubenswrapper[4692]: W0309 09:20:03.242838 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.181:6443: connect: connection refused Mar 09 09:20:03 crc kubenswrapper[4692]: E0309 09:20:03.242950 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.181:6443: connect: connection refused" logger="UnhandledError" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.270471 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.272386 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.272452 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.272463 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.272513 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:03 crc kubenswrapper[4692]: E0309 09:20:03.273339 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.181:6443: connect: connection refused" node="crc" Mar 09 09:20:03 crc kubenswrapper[4692]: I0309 09:20:03.794214 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.128659 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9e85d19fb172a4601c1dc0c88ef477a22e5b87c4bb1b34cf4036071368cb3c71"} Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.128853 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.130451 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.130523 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.130549 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.130834 4692 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="61345623a5480c63cfea6d85ec47c0251681da132d532aa6bfb508b8d5488349" exitCode=0 Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.130926 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.130933 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.130981 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.131016 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.131513 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"61345623a5480c63cfea6d85ec47c0251681da132d532aa6bfb508b8d5488349"} Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.131590 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.131822 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.131847 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.131855 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.131993 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132023 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132038 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132435 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132452 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132461 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132677 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132763 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:04 crc kubenswrapper[4692]: I0309 09:20:04.132783 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.136423 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7e82633557436e7cede06f97e5dd7e7fc475acff4f281840113f0a9c2f6169f6"} Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.136477 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9d24b074c00a7958169ac9e3a7b075cabbd2fffd2472efbabfe50f4b8c9c6371"} Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.136490 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"955c91ec4b3f4dfab9a76d69b01c4ce52515ba18b2be6d573574ca9c62fa93ca"} Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.136499 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2393ab2c9262269f0ffac03a6f7d6658cabd417080fa5fa3947a37ee99b4600f"} Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.136519 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.136604 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.137093 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.137307 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.137337 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.137352 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.137529 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.137547 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.137556 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.603654 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.603917 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.605376 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.605408 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.605418 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:05 crc kubenswrapper[4692]: I0309 09:20:05.751951 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.143518 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"de8c6b498f6d4920b8c93308a6ffe65bc7b6ba2d7ea438304d764d01218a205a"} Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.143621 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.143624 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.144949 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.144998 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.145026 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.145058 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.145061 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.145078 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.150087 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.150264 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.151195 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.151234 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.151247 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.156752 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.335292 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.474467 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.476154 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.476222 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.476234 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:06 crc kubenswrapper[4692]: I0309 09:20:06.476264 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.145663 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.145717 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.146310 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.147601 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.147677 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.147710 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.148784 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.148847 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.148872 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.149587 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.149625 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.149635 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.615581 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.706927 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:07 crc kubenswrapper[4692]: I0309 09:20:07.916345 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.149205 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.149263 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.149273 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.151413 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.151604 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.151738 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.151442 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.151921 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.151941 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.152319 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.152366 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:08 crc kubenswrapper[4692]: I0309 09:20:08.152378 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:09 crc kubenswrapper[4692]: I0309 09:20:09.872532 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 09 09:20:09 crc kubenswrapper[4692]: I0309 09:20:09.872834 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:09 crc kubenswrapper[4692]: I0309 09:20:09.874409 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:09 crc kubenswrapper[4692]: I0309 09:20:09.874454 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:09 crc kubenswrapper[4692]: I0309 09:20:09.874466 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:10 crc kubenswrapper[4692]: E0309 09:20:10.168891 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:20:10 crc kubenswrapper[4692]: I0309 09:20:10.708006 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 09:20:10 crc kubenswrapper[4692]: I0309 09:20:10.708119 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 09:20:13 crc kubenswrapper[4692]: I0309 09:20:13.800335 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:13 crc kubenswrapper[4692]: I0309 09:20:13.800440 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:13 crc kubenswrapper[4692]: I0309 09:20:13.804056 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:13 crc kubenswrapper[4692]: I0309 09:20:13.804158 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:13 crc kubenswrapper[4692]: I0309 09:20:13.804200 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.010485 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.012682 4692 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33810->192.168.126.11:17697: read: connection reset by peer" start-of-body= Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.012771 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33810->192.168.126.11:17697: read: connection reset by peer" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.041389 4692 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.041466 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.166227 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.168286 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9e85d19fb172a4601c1dc0c88ef477a22e5b87c4bb1b34cf4036071368cb3c71" exitCode=255 Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.168326 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9e85d19fb172a4601c1dc0c88ef477a22e5b87c4bb1b34cf4036071368cb3c71"} Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.168474 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.169446 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.169482 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.169494 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.170024 4692 scope.go:117] "RemoveContainer" containerID="9e85d19fb172a4601c1dc0c88ef477a22e5b87c4bb1b34cf4036071368cb3c71" Mar 09 09:20:14 crc kubenswrapper[4692]: W0309 09:20:14.199055 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 09 09:20:14 crc kubenswrapper[4692]: I0309 09:20:14.199216 4692 trace.go:236] Trace[996640855]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Mar-2026 09:20:04.197) (total time: 10001ms): Mar 09 09:20:14 crc kubenswrapper[4692]: Trace[996640855]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:20:14.199) Mar 09 09:20:14 crc kubenswrapper[4692]: Trace[996640855]: [10.00146126s] [10.00146126s] END Mar 09 09:20:14 crc kubenswrapper[4692]: E0309 09:20:14.199255 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.172916 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.175335 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23"} Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.175442 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.176145 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.176182 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.176193 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:15 crc kubenswrapper[4692]: E0309 09:20:15.338079 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 09 09:20:15 crc kubenswrapper[4692]: E0309 09:20:15.339748 4692 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.342610 4692 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.342666 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 09 09:20:15 crc kubenswrapper[4692]: E0309 09:20:15.343372 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 09:20:15 crc kubenswrapper[4692]: W0309 09:20:15.344668 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z Mar 09 09:20:15 crc kubenswrapper[4692]: E0309 09:20:15.344738 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.346868 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z Mar 09 09:20:15 crc kubenswrapper[4692]: W0309 09:20:15.348888 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z Mar 09 09:20:15 crc kubenswrapper[4692]: E0309 09:20:15.348959 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.350042 4692 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.350100 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 09 09:20:15 crc kubenswrapper[4692]: W0309 09:20:15.352875 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z Mar 09 09:20:15 crc kubenswrapper[4692]: E0309 09:20:15.352952 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:15 crc kubenswrapper[4692]: E0309 09:20:15.353546 4692 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:15Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b21c312144c93 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,LastTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.757202 4692 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]log ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]etcd ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/openshift.io-api-request-count-filter ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/openshift.io-startkubeinformers ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/generic-apiserver-start-informers ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/priority-and-fairness-config-consumer ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/priority-and-fairness-filter ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-apiextensions-informers ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-apiextensions-controllers ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/crd-informer-synced ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-system-namespaces-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-cluster-authentication-info-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-legacy-token-tracking-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-service-ip-repair-controllers ok Mar 09 09:20:15 crc kubenswrapper[4692]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Mar 09 09:20:15 crc kubenswrapper[4692]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/priority-and-fairness-config-producer ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/bootstrap-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/start-kube-aggregator-informers ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/apiservice-status-local-available-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/apiservice-status-remote-available-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/apiservice-registration-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/apiservice-wait-for-first-sync ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/apiservice-discovery-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/kube-apiserver-autoregistration ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]autoregister-completion ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/apiservice-openapi-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: [+]poststarthook/apiservice-openapiv3-controller ok Mar 09 09:20:15 crc kubenswrapper[4692]: livez check failed Mar 09 09:20:15 crc kubenswrapper[4692]: I0309 09:20:15.757262 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.013537 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:16Z is after 2026-02-23T05:33:13Z Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.179455 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.179972 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.181848 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" exitCode=255 Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.181884 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23"} Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.181918 4692 scope.go:117] "RemoveContainer" containerID="9e85d19fb172a4601c1dc0c88ef477a22e5b87c4bb1b34cf4036071368cb3c71" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.182044 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.183016 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.183045 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.183058 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:16 crc kubenswrapper[4692]: I0309 09:20:16.183510 4692 scope.go:117] "RemoveContainer" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" Mar 09 09:20:16 crc kubenswrapper[4692]: E0309 09:20:16.183669 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.011223 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:17Z is after 2026-02-23T05:33:13Z Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.188389 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 09:20:17 crc kubenswrapper[4692]: W0309 09:20:17.560534 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:17Z is after 2026-02-23T05:33:13Z Mar 09 09:20:17 crc kubenswrapper[4692]: E0309 09:20:17.560633 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:17Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.943146 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.943328 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.944876 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.944930 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.944941 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:17 crc kubenswrapper[4692]: I0309 09:20:17.958947 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 09 09:20:18 crc kubenswrapper[4692]: I0309 09:20:18.014045 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:18Z is after 2026-02-23T05:33:13Z Mar 09 09:20:18 crc kubenswrapper[4692]: I0309 09:20:18.195122 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:18 crc kubenswrapper[4692]: I0309 09:20:18.196702 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:18 crc kubenswrapper[4692]: I0309 09:20:18.196799 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:18 crc kubenswrapper[4692]: I0309 09:20:18.196822 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:19 crc kubenswrapper[4692]: I0309 09:20:19.013003 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:19Z is after 2026-02-23T05:33:13Z Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.013065 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:20Z is after 2026-02-23T05:33:13Z Mar 09 09:20:20 crc kubenswrapper[4692]: E0309 09:20:20.169671 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.708395 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.708467 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.757841 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.758036 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.759431 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.759480 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.759494 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.760086 4692 scope.go:117] "RemoveContainer" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" Mar 09 09:20:20 crc kubenswrapper[4692]: E0309 09:20:20.760295 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:20 crc kubenswrapper[4692]: I0309 09:20:20.762460 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.011440 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:21Z is after 2026-02-23T05:33:13Z Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.201102 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.201909 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.201935 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.201944 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.202437 4692 scope.go:117] "RemoveContainer" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" Mar 09 09:20:21 crc kubenswrapper[4692]: E0309 09:20:21.202618 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:21 crc kubenswrapper[4692]: E0309 09:20:21.741399 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:21Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.744474 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.745709 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.745751 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.745761 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:21 crc kubenswrapper[4692]: I0309 09:20:21.745793 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:21 crc kubenswrapper[4692]: E0309 09:20:21.748425 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:21Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 09:20:22 crc kubenswrapper[4692]: I0309 09:20:22.012265 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:22Z is after 2026-02-23T05:33:13Z Mar 09 09:20:22 crc kubenswrapper[4692]: W0309 09:20:22.825500 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:22Z is after 2026-02-23T05:33:13Z Mar 09 09:20:22 crc kubenswrapper[4692]: E0309 09:20:22.825625 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.011136 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:23Z is after 2026-02-23T05:33:13Z Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.372891 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 09:20:23 crc kubenswrapper[4692]: E0309 09:20:23.376115 4692 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:23Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.754066 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.754330 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.755827 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.755871 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.755884 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:23 crc kubenswrapper[4692]: I0309 09:20:23.756543 4692 scope.go:117] "RemoveContainer" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" Mar 09 09:20:23 crc kubenswrapper[4692]: E0309 09:20:23.756741 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:24 crc kubenswrapper[4692]: I0309 09:20:24.015212 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:24Z is after 2026-02-23T05:33:13Z Mar 09 09:20:24 crc kubenswrapper[4692]: I0309 09:20:24.040925 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:24 crc kubenswrapper[4692]: I0309 09:20:24.208822 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:24 crc kubenswrapper[4692]: I0309 09:20:24.209550 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:24 crc kubenswrapper[4692]: I0309 09:20:24.209587 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:24 crc kubenswrapper[4692]: I0309 09:20:24.209600 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:24 crc kubenswrapper[4692]: I0309 09:20:24.210115 4692 scope.go:117] "RemoveContainer" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" Mar 09 09:20:24 crc kubenswrapper[4692]: E0309 09:20:24.210304 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:24 crc kubenswrapper[4692]: W0309 09:20:24.565033 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:24Z is after 2026-02-23T05:33:13Z Mar 09 09:20:24 crc kubenswrapper[4692]: E0309 09:20:24.565094 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:25 crc kubenswrapper[4692]: I0309 09:20:25.013567 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:25Z is after 2026-02-23T05:33:13Z Mar 09 09:20:25 crc kubenswrapper[4692]: E0309 09:20:25.360378 4692 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:25Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b21c312144c93 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,LastTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:25 crc kubenswrapper[4692]: W0309 09:20:25.380554 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:25Z is after 2026-02-23T05:33:13Z Mar 09 09:20:25 crc kubenswrapper[4692]: E0309 09:20:25.380740 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:25Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:26 crc kubenswrapper[4692]: I0309 09:20:26.012667 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:26Z is after 2026-02-23T05:33:13Z Mar 09 09:20:26 crc kubenswrapper[4692]: W0309 09:20:26.076001 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:26Z is after 2026-02-23T05:33:13Z Mar 09 09:20:26 crc kubenswrapper[4692]: E0309 09:20:26.076073 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:27 crc kubenswrapper[4692]: I0309 09:20:27.013150 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:27Z is after 2026-02-23T05:33:13Z Mar 09 09:20:28 crc kubenswrapper[4692]: I0309 09:20:28.013838 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:28Z is after 2026-02-23T05:33:13Z Mar 09 09:20:28 crc kubenswrapper[4692]: E0309 09:20:28.745138 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:28Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 09:20:28 crc kubenswrapper[4692]: I0309 09:20:28.749446 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:28 crc kubenswrapper[4692]: I0309 09:20:28.750834 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:28 crc kubenswrapper[4692]: I0309 09:20:28.750931 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:28 crc kubenswrapper[4692]: I0309 09:20:28.750951 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:28 crc kubenswrapper[4692]: I0309 09:20:28.750979 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:28 crc kubenswrapper[4692]: E0309 09:20:28.754109 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:28Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 09:20:29 crc kubenswrapper[4692]: I0309 09:20:29.011724 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:29Z is after 2026-02-23T05:33:13Z Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.014024 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:30Z is after 2026-02-23T05:33:13Z Mar 09 09:20:30 crc kubenswrapper[4692]: E0309 09:20:30.169799 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.707349 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.707512 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.707622 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.707896 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.710011 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.710085 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.710103 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.710895 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"c3b766af9a91908ad63a850d6e36a104fb38f31802759317db0db0a364140055"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 09 09:20:30 crc kubenswrapper[4692]: I0309 09:20:30.711145 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://c3b766af9a91908ad63a850d6e36a104fb38f31802759317db0db0a364140055" gracePeriod=30 Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.011320 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:31Z is after 2026-02-23T05:33:13Z Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.228209 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.229009 4692 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c3b766af9a91908ad63a850d6e36a104fb38f31802759317db0db0a364140055" exitCode=255 Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.229056 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c3b766af9a91908ad63a850d6e36a104fb38f31802759317db0db0a364140055"} Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.229086 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ec43ca0d58f33a2866a5443e754b07798ac9ac2cc8d9e18822964f5a40c1fe67"} Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.229212 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.230077 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.230112 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:31 crc kubenswrapper[4692]: I0309 09:20:31.230123 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:32 crc kubenswrapper[4692]: I0309 09:20:32.011402 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:32Z is after 2026-02-23T05:33:13Z Mar 09 09:20:33 crc kubenswrapper[4692]: I0309 09:20:33.012536 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:33Z is after 2026-02-23T05:33:13Z Mar 09 09:20:34 crc kubenswrapper[4692]: I0309 09:20:34.012068 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:34Z is after 2026-02-23T05:33:13Z Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.013479 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:35Z is after 2026-02-23T05:33:13Z Mar 09 09:20:35 crc kubenswrapper[4692]: E0309 09:20:35.364001 4692 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:35Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b21c312144c93 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,LastTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.604338 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.604528 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.605496 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.605536 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.605544 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:35 crc kubenswrapper[4692]: E0309 09:20:35.748800 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:35Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.755122 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.756495 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.756538 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.756553 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:35 crc kubenswrapper[4692]: I0309 09:20:35.756582 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:35 crc kubenswrapper[4692]: E0309 09:20:35.759842 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:35Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 09:20:36 crc kubenswrapper[4692]: I0309 09:20:36.013832 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:36Z is after 2026-02-23T05:33:13Z Mar 09 09:20:36 crc kubenswrapper[4692]: I0309 09:20:36.071497 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:36 crc kubenswrapper[4692]: I0309 09:20:36.072677 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:36 crc kubenswrapper[4692]: I0309 09:20:36.072740 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:36 crc kubenswrapper[4692]: I0309 09:20:36.072758 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:36 crc kubenswrapper[4692]: I0309 09:20:36.073615 4692 scope.go:117] "RemoveContainer" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.011480 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:37Z is after 2026-02-23T05:33:13Z Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.242016 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.243985 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823"} Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.244148 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.244865 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.244919 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.244942 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.707130 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.707459 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.708940 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.708990 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:37 crc kubenswrapper[4692]: I0309 09:20:37.709003 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.011934 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:38Z is after 2026-02-23T05:33:13Z Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.247962 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.248426 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.250533 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823" exitCode=255 Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.250566 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823"} Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.250604 4692 scope.go:117] "RemoveContainer" containerID="0db07a3b596b5b3225d55b9bbe356933674656cd5fd68434a32ff2c115c9bd23" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.250794 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.252070 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.252109 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.252125 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:38 crc kubenswrapper[4692]: I0309 09:20:38.253069 4692 scope.go:117] "RemoveContainer" containerID="adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823" Mar 09 09:20:38 crc kubenswrapper[4692]: E0309 09:20:38.253381 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:38 crc kubenswrapper[4692]: W0309 09:20:38.618543 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:38Z is after 2026-02-23T05:33:13Z Mar 09 09:20:38 crc kubenswrapper[4692]: E0309 09:20:38.618625 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:38Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:39 crc kubenswrapper[4692]: I0309 09:20:39.014571 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:39Z is after 2026-02-23T05:33:13Z Mar 09 09:20:39 crc kubenswrapper[4692]: I0309 09:20:39.255019 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 09:20:39 crc kubenswrapper[4692]: I0309 09:20:39.868632 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 09:20:39 crc kubenswrapper[4692]: E0309 09:20:39.873733 4692 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:39Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:39 crc kubenswrapper[4692]: E0309 09:20:39.874933 4692 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 09 09:20:40 crc kubenswrapper[4692]: I0309 09:20:40.012359 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:40Z is after 2026-02-23T05:33:13Z Mar 09 09:20:40 crc kubenswrapper[4692]: E0309 09:20:40.169943 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:20:40 crc kubenswrapper[4692]: W0309 09:20:40.185462 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:40Z is after 2026-02-23T05:33:13Z Mar 09 09:20:40 crc kubenswrapper[4692]: E0309 09:20:40.185532 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:40Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:40 crc kubenswrapper[4692]: I0309 09:20:40.707963 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 09:20:40 crc kubenswrapper[4692]: I0309 09:20:40.708104 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 09:20:41 crc kubenswrapper[4692]: I0309 09:20:41.012345 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:41Z is after 2026-02-23T05:33:13Z Mar 09 09:20:42 crc kubenswrapper[4692]: I0309 09:20:42.011959 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:42Z is after 2026-02-23T05:33:13Z Mar 09 09:20:42 crc kubenswrapper[4692]: E0309 09:20:42.752677 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:42Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 09:20:42 crc kubenswrapper[4692]: I0309 09:20:42.760875 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:42 crc kubenswrapper[4692]: I0309 09:20:42.762213 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:42 crc kubenswrapper[4692]: I0309 09:20:42.762250 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:42 crc kubenswrapper[4692]: I0309 09:20:42.762261 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:42 crc kubenswrapper[4692]: I0309 09:20:42.762280 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:42 crc kubenswrapper[4692]: E0309 09:20:42.765187 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:42Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 09:20:43 crc kubenswrapper[4692]: I0309 09:20:43.012265 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:43Z is after 2026-02-23T05:33:13Z Mar 09 09:20:43 crc kubenswrapper[4692]: I0309 09:20:43.753606 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:43 crc kubenswrapper[4692]: I0309 09:20:43.753861 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:43 crc kubenswrapper[4692]: I0309 09:20:43.755319 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:43 crc kubenswrapper[4692]: I0309 09:20:43.755366 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:43 crc kubenswrapper[4692]: I0309 09:20:43.755375 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:43 crc kubenswrapper[4692]: I0309 09:20:43.755880 4692 scope.go:117] "RemoveContainer" containerID="adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823" Mar 09 09:20:43 crc kubenswrapper[4692]: E0309 09:20:43.756062 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:44 crc kubenswrapper[4692]: I0309 09:20:44.012204 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:44Z is after 2026-02-23T05:33:13Z Mar 09 09:20:44 crc kubenswrapper[4692]: I0309 09:20:44.041308 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:20:44 crc kubenswrapper[4692]: I0309 09:20:44.271684 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:44 crc kubenswrapper[4692]: I0309 09:20:44.273057 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:44 crc kubenswrapper[4692]: I0309 09:20:44.273121 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:44 crc kubenswrapper[4692]: I0309 09:20:44.273144 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:44 crc kubenswrapper[4692]: I0309 09:20:44.274245 4692 scope.go:117] "RemoveContainer" containerID="adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823" Mar 09 09:20:44 crc kubenswrapper[4692]: E0309 09:20:44.274608 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:45 crc kubenswrapper[4692]: I0309 09:20:45.014009 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:45Z is after 2026-02-23T05:33:13Z Mar 09 09:20:45 crc kubenswrapper[4692]: E0309 09:20:45.367999 4692 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:45Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189b21c312144c93 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,LastTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:46 crc kubenswrapper[4692]: I0309 09:20:46.012127 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:46Z is after 2026-02-23T05:33:13Z Mar 09 09:20:47 crc kubenswrapper[4692]: I0309 09:20:47.013461 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:47Z is after 2026-02-23T05:33:13Z Mar 09 09:20:47 crc kubenswrapper[4692]: W0309 09:20:47.267147 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:47Z is after 2026-02-23T05:33:13Z Mar 09 09:20:47 crc kubenswrapper[4692]: E0309 09:20:47.267245 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:48 crc kubenswrapper[4692]: I0309 09:20:48.012580 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:48Z is after 2026-02-23T05:33:13Z Mar 09 09:20:48 crc kubenswrapper[4692]: W0309 09:20:48.745253 4692 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:48Z is after 2026-02-23T05:33:13Z Mar 09 09:20:48 crc kubenswrapper[4692]: E0309 09:20:48.745383 4692 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:48Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 09 09:20:49 crc kubenswrapper[4692]: I0309 09:20:49.015516 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:49Z is after 2026-02-23T05:33:13Z Mar 09 09:20:49 crc kubenswrapper[4692]: E0309 09:20:49.758625 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:49Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 09 09:20:49 crc kubenswrapper[4692]: I0309 09:20:49.765774 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:49 crc kubenswrapper[4692]: I0309 09:20:49.767410 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:49 crc kubenswrapper[4692]: I0309 09:20:49.767450 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:49 crc kubenswrapper[4692]: I0309 09:20:49.767461 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:49 crc kubenswrapper[4692]: I0309 09:20:49.767483 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:49 crc kubenswrapper[4692]: E0309 09:20:49.772373 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:49Z is after 2026-02-23T05:33:13Z" node="crc" Mar 09 09:20:50 crc kubenswrapper[4692]: I0309 09:20:50.014298 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:50Z is after 2026-02-23T05:33:13Z Mar 09 09:20:50 crc kubenswrapper[4692]: E0309 09:20:50.170078 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:20:50 crc kubenswrapper[4692]: I0309 09:20:50.708691 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 09:20:50 crc kubenswrapper[4692]: I0309 09:20:50.708754 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 09:20:51 crc kubenswrapper[4692]: I0309 09:20:51.012327 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:51Z is after 2026-02-23T05:33:13Z Mar 09 09:20:52 crc kubenswrapper[4692]: I0309 09:20:52.012277 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:52Z is after 2026-02-23T05:33:13Z Mar 09 09:20:52 crc kubenswrapper[4692]: I0309 09:20:52.283635 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 09 09:20:52 crc kubenswrapper[4692]: I0309 09:20:52.283973 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:52 crc kubenswrapper[4692]: I0309 09:20:52.285599 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:52 crc kubenswrapper[4692]: I0309 09:20:52.285672 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:52 crc kubenswrapper[4692]: I0309 09:20:52.285691 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:53 crc kubenswrapper[4692]: I0309 09:20:53.013704 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-09T09:20:53Z is after 2026-02-23T05:33:13Z Mar 09 09:20:54 crc kubenswrapper[4692]: I0309 09:20:54.013990 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:20:55 crc kubenswrapper[4692]: I0309 09:20:55.012876 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:20:55 crc kubenswrapper[4692]: I0309 09:20:55.071083 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:55 crc kubenswrapper[4692]: I0309 09:20:55.072327 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:55 crc kubenswrapper[4692]: I0309 09:20:55.072393 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:55 crc kubenswrapper[4692]: I0309 09:20:55.072411 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:55 crc kubenswrapper[4692]: I0309 09:20:55.073260 4692 scope.go:117] "RemoveContainer" containerID="adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.073545 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.372320 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c312144c93 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,LastTimestamp:2026-03-09 09:20:00.006081683 +0000 UTC m=+0.830817264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.375738 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.379210 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.382711 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546de0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,LastTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.386212 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31b5eee3f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.161967679 +0000 UTC m=+0.986703270,LastTimestamp:2026-03-09 09:20:00.161967679 +0000 UTC m=+0.986703270,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.389881 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c3154652d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.173113821 +0000 UTC m=+0.997849402,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.394076 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546b24c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.173130061 +0000 UTC m=+0.997865632,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.397931 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546de0c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546de0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,LastTimestamp:2026-03-09 09:20:00.173141362 +0000 UTC m=+0.997876933,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.402123 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c3154652d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.17628832 +0000 UTC m=+1.001023901,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.406972 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546b24c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.17630844 +0000 UTC m=+1.001044021,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.410783 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546de0c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546de0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,LastTimestamp:2026-03-09 09:20:00.17631736 +0000 UTC m=+1.001052941,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.415890 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c3154652d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.176432024 +0000 UTC m=+1.001167605,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.419053 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546b24c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.176472665 +0000 UTC m=+1.001208246,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.422846 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546de0c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546de0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,LastTimestamp:2026-03-09 09:20:00.176481985 +0000 UTC m=+1.001217566,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.426973 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c3154652d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.177446012 +0000 UTC m=+1.002181583,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.431180 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546b24c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.177464552 +0000 UTC m=+1.002200133,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.435411 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546de0c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546de0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,LastTimestamp:2026-03-09 09:20:00.177474213 +0000 UTC m=+1.002209794,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.439987 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c3154652d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.177651518 +0000 UTC m=+1.002387099,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.444974 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546b24c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.177664988 +0000 UTC m=+1.002400569,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.448564 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546de0c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546de0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,LastTimestamp:2026-03-09 09:20:00.177674728 +0000 UTC m=+1.002410299,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.452334 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c3154652d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.178967464 +0000 UTC m=+1.003703055,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.455776 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546b24c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.179000555 +0000 UTC m=+1.003736146,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.459277 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546de0c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546de0c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059727372 +0000 UTC m=+0.884462973,LastTimestamp:2026-03-09 09:20:00.179014626 +0000 UTC m=+1.003750217,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.462787 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c3154652d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c3154652d3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059691731 +0000 UTC m=+0.884427322,LastTimestamp:2026-03-09 09:20:00.179289253 +0000 UTC m=+1.004024824,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.465921 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189b21c31546b24c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189b21c31546b24c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.059716172 +0000 UTC m=+0.884451763,LastTimestamp:2026-03-09 09:20:00.179304454 +0000 UTC m=+1.004040035,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.470358 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c3348b2c78 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.584297592 +0000 UTC m=+1.409033173,LastTimestamp:2026-03-09 09:20:00.584297592 +0000 UTC m=+1.409033173,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.473950 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b21c3348c4416 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.584369174 +0000 UTC m=+1.409104765,LastTimestamp:2026-03-09 09:20:00.584369174 +0000 UTC m=+1.409104765,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.477317 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c334d74978 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.589285752 +0000 UTC m=+1.414021353,LastTimestamp:2026-03-09 09:20:00.589285752 +0000 UTC m=+1.414021353,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.482373 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c335317a97 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.595196567 +0000 UTC m=+1.419932148,LastTimestamp:2026-03-09 09:20:00.595196567 +0000 UTC m=+1.419932148,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.485916 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c335866fad openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:00.600764333 +0000 UTC m=+1.425499944,LastTimestamp:2026-03-09 09:20:00.600764333 +0000 UTC m=+1.425499944,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.490047 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c355a6cfff openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.139757055 +0000 UTC m=+1.964492636,LastTimestamp:2026-03-09 09:20:01.139757055 +0000 UTC m=+1.964492636,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.493503 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c355a7f5ad openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.139832237 +0000 UTC m=+1.964567828,LastTimestamp:2026-03-09 09:20:01.139832237 +0000 UTC m=+1.964567828,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.496664 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b21c355aa3d49 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.139981641 +0000 UTC m=+1.964717222,LastTimestamp:2026-03-09 09:20:01.139981641 +0000 UTC m=+1.964717222,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.499800 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c355ab1941 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.140037953 +0000 UTC m=+1.964773544,LastTimestamp:2026-03-09 09:20:01.140037953 +0000 UTC m=+1.964773544,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.503997 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c35679c196 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.153581462 +0000 UTC m=+1.978317043,LastTimestamp:2026-03-09 09:20:01.153581462 +0000 UTC m=+1.978317043,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.507581 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c356b23ba5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.157282725 +0000 UTC m=+1.982018306,LastTimestamp:2026-03-09 09:20:01.157282725 +0000 UTC m=+1.982018306,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.510739 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c356c45e96 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.158471318 +0000 UTC m=+1.983206899,LastTimestamp:2026-03-09 09:20:01.158471318 +0000 UTC m=+1.983206899,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.513658 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b21c356c48a24 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.158482468 +0000 UTC m=+1.983218049,LastTimestamp:2026-03-09 09:20:01.158482468 +0000 UTC m=+1.983218049,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.517949 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c356c46652 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.158473298 +0000 UTC m=+1.983208889,LastTimestamp:2026-03-09 09:20:01.158473298 +0000 UTC m=+1.983208889,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.521018 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c356d628db openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.159637211 +0000 UTC m=+1.984372802,LastTimestamp:2026-03-09 09:20:01.159637211 +0000 UTC m=+1.984372802,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.524298 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c3580137b6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.179236278 +0000 UTC m=+2.003971879,LastTimestamp:2026-03-09 09:20:01.179236278 +0000 UTC m=+2.003971879,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.528114 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c367e1d717 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.445615383 +0000 UTC m=+2.270351004,LastTimestamp:2026-03-09 09:20:01.445615383 +0000 UTC m=+2.270351004,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.531716 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c36877b8ad openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.455437997 +0000 UTC m=+2.280173578,LastTimestamp:2026-03-09 09:20:01.455437997 +0000 UTC m=+2.280173578,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.535446 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c3688f6c45 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.456991301 +0000 UTC m=+2.281726882,LastTimestamp:2026-03-09 09:20:01.456991301 +0000 UTC m=+2.281726882,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.538872 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c3733790ab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.635782827 +0000 UTC m=+2.460518408,LastTimestamp:2026-03-09 09:20:01.635782827 +0000 UTC m=+2.460518408,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.542871 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c3741ecc45 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.650936901 +0000 UTC m=+2.475672472,LastTimestamp:2026-03-09 09:20:01.650936901 +0000 UTC m=+2.475672472,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.547423 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c3742f67df openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.652025311 +0000 UTC m=+2.476760892,LastTimestamp:2026-03-09 09:20:01.652025311 +0000 UTC m=+2.476760892,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.551796 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c381059a6e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.86738955 +0000 UTC m=+2.692125131,LastTimestamp:2026-03-09 09:20:01.86738955 +0000 UTC m=+2.692125131,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.556026 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c3820bda36 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.88457631 +0000 UTC m=+2.709311891,LastTimestamp:2026-03-09 09:20:01.88457631 +0000 UTC m=+2.709311891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.560982 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c38e6bd55b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.092193115 +0000 UTC m=+2.916928696,LastTimestamp:2026-03-09 09:20:02.092193115 +0000 UTC m=+2.916928696,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.565047 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b21c38eb047ae openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.09667883 +0000 UTC m=+2.921414441,LastTimestamp:2026-03-09 09:20:02.09667883 +0000 UTC m=+2.921414441,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.571278 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c38ece1a41 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.098633281 +0000 UTC m=+2.923368882,LastTimestamp:2026-03-09 09:20:02.098633281 +0000 UTC m=+2.923368882,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.576256 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c38f5dcfa7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.108051367 +0000 UTC m=+2.932786948,LastTimestamp:2026-03-09 09:20:02.108051367 +0000 UTC m=+2.932786948,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.581366 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c39b37e0af openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.306891951 +0000 UTC m=+3.131627542,LastTimestamp:2026-03-09 09:20:02.306891951 +0000 UTC m=+3.131627542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.585316 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b21c39bb464f8 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.31505228 +0000 UTC m=+3.139787861,LastTimestamp:2026-03-09 09:20:02.31505228 +0000 UTC m=+3.139787861,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.589437 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c39bd17958 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.31695804 +0000 UTC m=+3.141693621,LastTimestamp:2026-03-09 09:20:02.31695804 +0000 UTC m=+3.141693621,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.593419 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c39bedcae8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.318813928 +0000 UTC m=+3.143549509,LastTimestamp:2026-03-09 09:20:02.318813928 +0000 UTC m=+3.143549509,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.598339 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c39bfdaa9c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.319854236 +0000 UTC m=+3.144589817,LastTimestamp:2026-03-09 09:20:02.319854236 +0000 UTC m=+3.144589817,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.603153 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c39bfe4741 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.319894337 +0000 UTC m=+3.144629918,LastTimestamp:2026-03-09 09:20:02.319894337 +0000 UTC m=+3.144629918,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.607948 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189b21c39ca2c9d8 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.330675672 +0000 UTC m=+3.155411244,LastTimestamp:2026-03-09 09:20:02.330675672 +0000 UTC m=+3.155411244,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.612560 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c39cf2db5b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.335923035 +0000 UTC m=+3.160658616,LastTimestamp:2026-03-09 09:20:02.335923035 +0000 UTC m=+3.160658616,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.616271 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c39d3b997b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.340690299 +0000 UTC m=+3.165425870,LastTimestamp:2026-03-09 09:20:02.340690299 +0000 UTC m=+3.165425870,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.620637 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c39d4ab278 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.341679736 +0000 UTC m=+3.166415327,LastTimestamp:2026-03-09 09:20:02.341679736 +0000 UTC m=+3.166415327,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.623936 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c3a77beb77 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.512677751 +0000 UTC m=+3.337413322,LastTimestamp:2026-03-09 09:20:02.512677751 +0000 UTC m=+3.337413322,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.628530 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3a7937b36 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.514221878 +0000 UTC m=+3.338957459,LastTimestamp:2026-03-09 09:20:02.514221878 +0000 UTC m=+3.338957459,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.633106 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3a83c5189 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.525286793 +0000 UTC m=+3.350022374,LastTimestamp:2026-03-09 09:20:02.525286793 +0000 UTC m=+3.350022374,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.639813 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3a84e648a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.526471306 +0000 UTC m=+3.351206887,LastTimestamp:2026-03-09 09:20:02.526471306 +0000 UTC m=+3.351206887,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.643943 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c3a88187e1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.529822689 +0000 UTC m=+3.354558270,LastTimestamp:2026-03-09 09:20:02.529822689 +0000 UTC m=+3.354558270,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.647829 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c3a8905501 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.530792705 +0000 UTC m=+3.355528286,LastTimestamp:2026-03-09 09:20:02.530792705 +0000 UTC m=+3.355528286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.650845 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c3b56253f2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.745881586 +0000 UTC m=+3.570617177,LastTimestamp:2026-03-09 09:20:02.745881586 +0000 UTC m=+3.570617177,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.654234 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3b5804c2e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.747845678 +0000 UTC m=+3.572581269,LastTimestamp:2026-03-09 09:20:02.747845678 +0000 UTC m=+3.572581269,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.657606 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3b65ce0a7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.762301607 +0000 UTC m=+3.587037178,LastTimestamp:2026-03-09 09:20:02.762301607 +0000 UTC m=+3.587037178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.661655 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3b66d4f37 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.763378487 +0000 UTC m=+3.588114078,LastTimestamp:2026-03-09 09:20:02.763378487 +0000 UTC m=+3.588114078,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.666289 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189b21c3b6809237 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.764640823 +0000 UTC m=+3.589376404,LastTimestamp:2026-03-09 09:20:02.764640823 +0000 UTC m=+3.589376404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.670944 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3c07fb74a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.932356938 +0000 UTC m=+3.757092519,LastTimestamp:2026-03-09 09:20:02.932356938 +0000 UTC m=+3.757092519,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.674941 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3c14539cf openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.945300943 +0000 UTC m=+3.770036524,LastTimestamp:2026-03-09 09:20:02.945300943 +0000 UTC m=+3.770036524,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.677590 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3c159c362 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.946646882 +0000 UTC m=+3.771382463,LastTimestamp:2026-03-09 09:20:02.946646882 +0000 UTC m=+3.771382463,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.682063 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c3cbb51844 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:03.120404548 +0000 UTC m=+3.945140139,LastTimestamp:2026-03-09 09:20:03.120404548 +0000 UTC m=+3.945140139,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.687859 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3ce13a2d3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:03.160154835 +0000 UTC m=+3.984890426,LastTimestamp:2026-03-09 09:20:03.160154835 +0000 UTC m=+3.984890426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.693053 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3cf32175b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:03.178927963 +0000 UTC m=+4.003663544,LastTimestamp:2026-03-09 09:20:03.178927963 +0000 UTC m=+4.003663544,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.697724 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c3d6876850 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:03.30195976 +0000 UTC m=+4.126695342,LastTimestamp:2026-03-09 09:20:03.30195976 +0000 UTC m=+4.126695342,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.701858 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c3d7625fb0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:03.316309936 +0000 UTC m=+4.141045527,LastTimestamp:2026-03-09 09:20:03.316309936 +0000 UTC m=+4.141045527,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.707687 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c4080d760c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.132828684 +0000 UTC m=+4.957564265,LastTimestamp:2026-03-09 09:20:04.132828684 +0000 UTC m=+4.957564265,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.712056 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c413e96f30 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.331794224 +0000 UTC m=+5.156529815,LastTimestamp:2026-03-09 09:20:04.331794224 +0000 UTC m=+5.156529815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.717480 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c41462874f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.339730255 +0000 UTC m=+5.164465846,LastTimestamp:2026-03-09 09:20:04.339730255 +0000 UTC m=+5.164465846,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.721006 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c41476b9f7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.341053943 +0000 UTC m=+5.165789534,LastTimestamp:2026-03-09 09:20:04.341053943 +0000 UTC m=+5.165789534,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.726355 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c41fc4a9d0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.530710992 +0000 UTC m=+5.355446573,LastTimestamp:2026-03-09 09:20:04.530710992 +0000 UTC m=+5.355446573,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.729647 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c420baef86 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.546850694 +0000 UTC m=+5.371586315,LastTimestamp:2026-03-09 09:20:04.546850694 +0000 UTC m=+5.371586315,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.735156 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c420ca6e9f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.547866271 +0000 UTC m=+5.372601882,LastTimestamp:2026-03-09 09:20:04.547866271 +0000 UTC m=+5.372601882,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.741708 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c42cf12686 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.75173031 +0000 UTC m=+5.576465931,LastTimestamp:2026-03-09 09:20:04.75173031 +0000 UTC m=+5.576465931,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.746549 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c42dbc1920 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.765030688 +0000 UTC m=+5.589766299,LastTimestamp:2026-03-09 09:20:04.765030688 +0000 UTC m=+5.589766299,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.752133 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c42dd75b27 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.766817063 +0000 UTC m=+5.591552664,LastTimestamp:2026-03-09 09:20:04.766817063 +0000 UTC m=+5.591552664,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.757213 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c439706da7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.961398183 +0000 UTC m=+5.786133764,LastTimestamp:2026-03-09 09:20:04.961398183 +0000 UTC m=+5.786133764,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.763747 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c43a22837c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.97306918 +0000 UTC m=+5.797804791,LastTimestamp:2026-03-09 09:20:04.97306918 +0000 UTC m=+5.797804791,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.770505 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c43a3548c7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:04.974299335 +0000 UTC m=+5.799034916,LastTimestamp:2026-03-09 09:20:04.974299335 +0000 UTC m=+5.799034916,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.776944 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c443ebdea4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:05.137260196 +0000 UTC m=+5.961995777,LastTimestamp:2026-03-09 09:20:05.137260196 +0000 UTC m=+5.961995777,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.784296 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189b21c44488b96f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:05.147539823 +0000 UTC m=+5.972275404,LastTimestamp:2026-03-09 09:20:05.147539823 +0000 UTC m=+5.972275404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.794449 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 09:20:55 crc kubenswrapper[4692]: &Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff7efcb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 09:20:55 crc kubenswrapper[4692]: body: Mar 09 09:20:55 crc kubenswrapper[4692]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708086731 +0000 UTC m=+11.532822312,LastTimestamp:2026-03-09 09:20:10.708086731 +0000 UTC m=+11.532822312,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 09:20:55 crc kubenswrapper[4692]: > Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.799889 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff9a8d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708199635 +0000 UTC m=+11.532935216,LastTimestamp:2026-03-09 09:20:10.708199635 +0000 UTC m=+11.532935216,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.806629 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 09 09:20:55 crc kubenswrapper[4692]: &Event{ObjectMeta:{kube-apiserver-crc.189b21c654f0f89a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": read tcp 192.168.126.11:33810->192.168.126.11:17697: read: connection reset by peer Mar 09 09:20:55 crc kubenswrapper[4692]: body: Mar 09 09:20:55 crc kubenswrapper[4692]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:14.012741786 +0000 UTC m=+14.837477367,LastTimestamp:2026-03-09 09:20:14.012741786 +0000 UTC m=+14.837477367,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 09:20:55 crc kubenswrapper[4692]: > Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.812761 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c654f2026e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:33810->192.168.126.11:17697: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:14.012809838 +0000 UTC m=+14.837545419,LastTimestamp:2026-03-09 09:20:14.012809838 +0000 UTC m=+14.837545419,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.819357 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 09 09:20:55 crc kubenswrapper[4692]: &Event{ObjectMeta:{kube-apiserver-crc.189b21c656a6f387 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 09 09:20:55 crc kubenswrapper[4692]: body: Mar 09 09:20:55 crc kubenswrapper[4692]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:14.041445255 +0000 UTC m=+14.866180836,LastTimestamp:2026-03-09 09:20:14.041445255 +0000 UTC m=+14.866180836,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 09:20:55 crc kubenswrapper[4692]: > Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.825205 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c656a7a49b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:14.041490587 +0000 UTC m=+14.866226168,LastTimestamp:2026-03-09 09:20:14.041490587 +0000 UTC m=+14.866226168,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.832323 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189b21c3c159c362\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3c159c362 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:02.946646882 +0000 UTC m=+3.771382463,LastTimestamp:2026-03-09 09:20:14.171205934 +0000 UTC m=+14.995941515,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.839023 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189b21c3ce13a2d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3ce13a2d3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:03.160154835 +0000 UTC m=+3.984890426,LastTimestamp:2026-03-09 09:20:14.357823793 +0000 UTC m=+15.182559374,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.844815 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189b21c3cf32175b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189b21c3cf32175b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:03.178927963 +0000 UTC m=+4.003663544,LastTimestamp:2026-03-09 09:20:14.366981501 +0000 UTC m=+15.191717082,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.850878 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c58ff7efcb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 09:20:55 crc kubenswrapper[4692]: &Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff7efcb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 09:20:55 crc kubenswrapper[4692]: body: Mar 09 09:20:55 crc kubenswrapper[4692]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708086731 +0000 UTC m=+11.532822312,LastTimestamp:2026-03-09 09:20:20.708449421 +0000 UTC m=+21.533185002,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 09:20:55 crc kubenswrapper[4692]: > Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.855479 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c58ff9a8d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff9a8d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708199635 +0000 UTC m=+11.532935216,LastTimestamp:2026-03-09 09:20:20.708534813 +0000 UTC m=+21.533270394,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.861384 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c58ff7efcb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 09:20:55 crc kubenswrapper[4692]: &Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff7efcb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 09:20:55 crc kubenswrapper[4692]: body: Mar 09 09:20:55 crc kubenswrapper[4692]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708086731 +0000 UTC m=+11.532822312,LastTimestamp:2026-03-09 09:20:30.707472641 +0000 UTC m=+31.532208252,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 09:20:55 crc kubenswrapper[4692]: > Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.865937 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c58ff9a8d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff9a8d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708199635 +0000 UTC m=+11.532935216,LastTimestamp:2026-03-09 09:20:30.707571314 +0000 UTC m=+31.532306935,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.870553 4692 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21ca383df2a4 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:30.711116452 +0000 UTC m=+31.535852073,LastTimestamp:2026-03-09 09:20:30.711116452 +0000 UTC m=+31.535852073,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.874428 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c356d628db\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c356d628db openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.159637211 +0000 UTC m=+1.984372802,LastTimestamp:2026-03-09 09:20:30.83130474 +0000 UTC m=+31.656040331,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.878984 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c367e1d717\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c367e1d717 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.445615383 +0000 UTC m=+2.270351004,LastTimestamp:2026-03-09 09:20:30.99120611 +0000 UTC m=+31.815941691,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.884301 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c36877b8ad\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c36877b8ad openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:01.455437997 +0000 UTC m=+2.280173578,LastTimestamp:2026-03-09 09:20:31.001091099 +0000 UTC m=+31.825826700,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.890863 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c58ff7efcb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 09:20:55 crc kubenswrapper[4692]: &Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff7efcb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 09:20:55 crc kubenswrapper[4692]: body: Mar 09 09:20:55 crc kubenswrapper[4692]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708086731 +0000 UTC m=+11.532822312,LastTimestamp:2026-03-09 09:20:40.708057637 +0000 UTC m=+41.532793298,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 09:20:55 crc kubenswrapper[4692]: > Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.895755 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c58ff9a8d3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff9a8d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708199635 +0000 UTC m=+11.532935216,LastTimestamp:2026-03-09 09:20:40.70815432 +0000 UTC m=+41.532889941,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:20:55 crc kubenswrapper[4692]: E0309 09:20:55.902243 4692 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189b21c58ff7efcb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 09 09:20:55 crc kubenswrapper[4692]: &Event{ObjectMeta:{kube-controller-manager-crc.189b21c58ff7efcb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 09 09:20:55 crc kubenswrapper[4692]: body: Mar 09 09:20:55 crc kubenswrapper[4692]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:20:10.708086731 +0000 UTC m=+11.532822312,LastTimestamp:2026-03-09 09:20:50.708736985 +0000 UTC m=+51.533472576,Count:5,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 09 09:20:55 crc kubenswrapper[4692]: > Mar 09 09:20:56 crc kubenswrapper[4692]: I0309 09:20:56.013407 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:20:56 crc kubenswrapper[4692]: E0309 09:20:56.768032 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 09 09:20:56 crc kubenswrapper[4692]: I0309 09:20:56.772628 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:20:56 crc kubenswrapper[4692]: I0309 09:20:56.774265 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:20:56 crc kubenswrapper[4692]: I0309 09:20:56.774538 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:20:56 crc kubenswrapper[4692]: I0309 09:20:56.774717 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:20:56 crc kubenswrapper[4692]: I0309 09:20:56.774880 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:20:56 crc kubenswrapper[4692]: E0309 09:20:56.782511 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 09 09:20:57 crc kubenswrapper[4692]: I0309 09:20:57.013577 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:20:58 crc kubenswrapper[4692]: I0309 09:20:58.014769 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:20:59 crc kubenswrapper[4692]: I0309 09:20:59.015413 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.016879 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:00 crc kubenswrapper[4692]: E0309 09:21:00.170406 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.707816 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.707902 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.707970 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.708196 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.709446 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.709471 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.709482 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.710205 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"ec43ca0d58f33a2866a5443e754b07798ac9ac2cc8d9e18822964f5a40c1fe67"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 09 09:21:00 crc kubenswrapper[4692]: I0309 09:21:00.710286 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://ec43ca0d58f33a2866a5443e754b07798ac9ac2cc8d9e18822964f5a40c1fe67" gracePeriod=30 Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.012779 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.314306 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.315215 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.315512 4692 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ec43ca0d58f33a2866a5443e754b07798ac9ac2cc8d9e18822964f5a40c1fe67" exitCode=255 Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.315546 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ec43ca0d58f33a2866a5443e754b07798ac9ac2cc8d9e18822964f5a40c1fe67"} Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.315585 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"241bf2d436536516117a99f299e37f923f8626babfa9bcab51e03a5c0198cf61"} Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.315602 4692 scope.go:117] "RemoveContainer" containerID="c3b766af9a91908ad63a850d6e36a104fb38f31802759317db0db0a364140055" Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.315719 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.317492 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.317527 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:01 crc kubenswrapper[4692]: I0309 09:21:01.317537 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:02 crc kubenswrapper[4692]: I0309 09:21:02.013948 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:02 crc kubenswrapper[4692]: I0309 09:21:02.319434 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 09:21:03 crc kubenswrapper[4692]: I0309 09:21:03.013825 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:03 crc kubenswrapper[4692]: E0309 09:21:03.773811 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 09 09:21:03 crc kubenswrapper[4692]: I0309 09:21:03.782848 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:03 crc kubenswrapper[4692]: I0309 09:21:03.784179 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:03 crc kubenswrapper[4692]: I0309 09:21:03.784213 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:03 crc kubenswrapper[4692]: I0309 09:21:03.784223 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:03 crc kubenswrapper[4692]: I0309 09:21:03.784245 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:21:03 crc kubenswrapper[4692]: E0309 09:21:03.788965 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 09 09:21:04 crc kubenswrapper[4692]: I0309 09:21:04.014456 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:05 crc kubenswrapper[4692]: I0309 09:21:05.013690 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:05 crc kubenswrapper[4692]: I0309 09:21:05.603679 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:21:05 crc kubenswrapper[4692]: I0309 09:21:05.603813 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:05 crc kubenswrapper[4692]: I0309 09:21:05.604702 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:05 crc kubenswrapper[4692]: I0309 09:21:05.604728 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:05 crc kubenswrapper[4692]: I0309 09:21:05.604738 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.012533 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.070989 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.071941 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.072000 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.072013 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.072724 4692 scope.go:117] "RemoveContainer" containerID="adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.330118 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.331749 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f"} Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.331882 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.332550 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.332578 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:06 crc kubenswrapper[4692]: I0309 09:21:06.332586 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.012051 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.336727 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.338051 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.339916 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" exitCode=255 Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.339994 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f"} Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.340145 4692 scope.go:117] "RemoveContainer" containerID="adaa21b5b4926d0f4929beda3897af7bf15ad9e885acb620f06c449968469823" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.340264 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.341304 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.341419 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.341542 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.342756 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:21:07 crc kubenswrapper[4692]: E0309 09:21:07.342983 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.707635 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.707989 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.708913 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.708939 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.708947 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:07 crc kubenswrapper[4692]: I0309 09:21:07.712775 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:21:08 crc kubenswrapper[4692]: I0309 09:21:08.011590 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:08 crc kubenswrapper[4692]: I0309 09:21:08.344976 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 09:21:08 crc kubenswrapper[4692]: I0309 09:21:08.346886 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:08 crc kubenswrapper[4692]: I0309 09:21:08.347560 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:08 crc kubenswrapper[4692]: I0309 09:21:08.347593 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:08 crc kubenswrapper[4692]: I0309 09:21:08.347617 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:09 crc kubenswrapper[4692]: I0309 09:21:09.014041 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:10 crc kubenswrapper[4692]: I0309 09:21:10.013540 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:10 crc kubenswrapper[4692]: E0309 09:21:10.170532 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:21:10 crc kubenswrapper[4692]: E0309 09:21:10.778790 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 09 09:21:10 crc kubenswrapper[4692]: I0309 09:21:10.789102 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:10 crc kubenswrapper[4692]: I0309 09:21:10.790127 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:10 crc kubenswrapper[4692]: I0309 09:21:10.790261 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:10 crc kubenswrapper[4692]: I0309 09:21:10.790287 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:10 crc kubenswrapper[4692]: I0309 09:21:10.790324 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:21:10 crc kubenswrapper[4692]: E0309 09:21:10.795064 4692 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 09 09:21:11 crc kubenswrapper[4692]: I0309 09:21:11.011828 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:11 crc kubenswrapper[4692]: I0309 09:21:11.876206 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 09 09:21:11 crc kubenswrapper[4692]: I0309 09:21:11.891893 4692 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 09 09:21:12 crc kubenswrapper[4692]: I0309 09:21:12.014598 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:13 crc kubenswrapper[4692]: I0309 09:21:13.012655 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:13 crc kubenswrapper[4692]: I0309 09:21:13.753979 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:21:13 crc kubenswrapper[4692]: I0309 09:21:13.754156 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:13 crc kubenswrapper[4692]: I0309 09:21:13.755505 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:13 crc kubenswrapper[4692]: I0309 09:21:13.755582 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:13 crc kubenswrapper[4692]: I0309 09:21:13.755603 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:13 crc kubenswrapper[4692]: I0309 09:21:13.756699 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:21:13 crc kubenswrapper[4692]: E0309 09:21:13.757026 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.014685 4692 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.041260 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.361999 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.363605 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.363664 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.363684 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.364824 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:21:14 crc kubenswrapper[4692]: E0309 09:21:14.365195 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.964371 4692 csr.go:261] certificate signing request csr-chspz is approved, waiting to be issued Mar 09 09:21:14 crc kubenswrapper[4692]: I0309 09:21:14.977778 4692 csr.go:257] certificate signing request csr-chspz is issued Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.036663 4692 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.609374 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.609588 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.610897 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.610995 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.611009 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.808337 4692 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.979249 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-09 02:34:45.433924907 +0000 UTC Mar 09 09:21:15 crc kubenswrapper[4692]: I0309 09:21:15.979394 4692 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7337h13m29.454542573s for next certificate rotation Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.071515 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.072938 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.073030 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.073051 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.795145 4692 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.797035 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.797108 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.797124 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.797286 4692 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.807825 4692 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.808366 4692 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.808410 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.812500 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.812546 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.812561 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.812584 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.812598 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:17Z","lastTransitionTime":"2026-03-09T09:21:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.828696 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.838298 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.838345 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.838360 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.838385 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.838402 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:17Z","lastTransitionTime":"2026-03-09T09:21:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.850782 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.859467 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.859510 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.859526 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.859546 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.859560 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:17Z","lastTransitionTime":"2026-03-09T09:21:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.872269 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.883855 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.883902 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.883913 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.883932 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:17 crc kubenswrapper[4692]: I0309 09:21:17.883945 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:17Z","lastTransitionTime":"2026-03-09T09:21:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.896647 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.896761 4692 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.896784 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:17 crc kubenswrapper[4692]: E0309 09:21:17.997569 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.098361 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.199207 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.300295 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.400926 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.501609 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.602385 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.702540 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.803183 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:18 crc kubenswrapper[4692]: E0309 09:21:18.903350 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.004330 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.105115 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.206087 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.306767 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.407301 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.508086 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.609136 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.710252 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.811352 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:19 crc kubenswrapper[4692]: E0309 09:21:19.912292 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.013026 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.113395 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.171459 4692 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.220055 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.321241 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.421887 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.522021 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.623054 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.724118 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.824645 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:20 crc kubenswrapper[4692]: E0309 09:21:20.925688 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:21 crc kubenswrapper[4692]: E0309 09:21:21.025878 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:21 crc kubenswrapper[4692]: E0309 09:21:21.126967 4692 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.170637 4692 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.229479 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.229559 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.229583 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.229613 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.229636 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.332313 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.332371 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.332383 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.332396 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.332408 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.435765 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.435836 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.435857 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.435888 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.435911 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.538659 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.538697 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.538705 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.538719 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.538728 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.641305 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.641367 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.641384 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.641407 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.641425 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.744272 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.744313 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.744323 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.744338 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.744349 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.847480 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.847525 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.847535 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.847552 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.847564 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.950312 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.950387 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.950396 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.950410 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:21 crc kubenswrapper[4692]: I0309 09:21:21.950419 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:21Z","lastTransitionTime":"2026-03-09T09:21:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.041640 4692 apiserver.go:52] "Watching apiserver" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.047029 4692 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.047534 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb"] Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.048003 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.048053 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.048107 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.048302 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.048327 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.048448 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.048472 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.048330 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.048557 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.049617 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.050152 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.050427 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.050600 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.050743 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.050868 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.051417 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.051716 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.052581 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.052579 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.052725 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.052757 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.052782 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.052801 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.070862 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.085939 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.099148 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.114604 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.116883 4692 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.128191 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.140249 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.152735 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.155342 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.155383 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.155393 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.155410 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.155422 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.165438 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.182866 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183037 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183153 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183330 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183399 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183433 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183580 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183623 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183656 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.183743 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184077 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184406 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184476 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184522 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184544 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184565 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184586 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184609 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184633 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184656 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184679 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184703 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184759 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184781 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184800 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184816 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184824 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184877 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184906 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184930 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184956 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184974 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.184995 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185014 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185016 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185034 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185060 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185082 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185105 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185126 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185149 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185189 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185209 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185229 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185251 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185272 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185293 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185314 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185332 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185354 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185395 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185373 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185439 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185458 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185477 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185499 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185516 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185620 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185640 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185657 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185673 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185719 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185749 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185782 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185806 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185829 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185853 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185875 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185900 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185922 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185946 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185974 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185998 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186048 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186067 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186085 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186101 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186129 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186177 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186203 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186230 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186257 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186286 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186309 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186359 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186384 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186408 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186427 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186444 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186465 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186487 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186508 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186533 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186554 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186575 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186665 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186689 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186712 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186740 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186764 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186785 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186803 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186819 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186834 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186853 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186870 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186889 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186906 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186923 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186938 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186954 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186969 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186989 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187006 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187022 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187038 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187053 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187070 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187086 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187103 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187129 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187150 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187185 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187202 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187217 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187237 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187254 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187273 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187291 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187309 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187326 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187343 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187398 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187423 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187446 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187471 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187498 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187524 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187551 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187574 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187592 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187610 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187627 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187645 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187663 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187682 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187715 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187739 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187756 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187773 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187797 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187814 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187833 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187852 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187870 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187889 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187908 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188129 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188253 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188279 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188301 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188323 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188345 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188366 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188387 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188410 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188433 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188456 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188476 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188499 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188524 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188556 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188580 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188607 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188630 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188654 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188681 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188706 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188760 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188788 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188814 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188841 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188866 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188895 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188925 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188949 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188974 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188996 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189022 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189046 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189068 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189092 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189117 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189142 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189473 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189929 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189956 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189991 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190015 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190035 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190059 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190082 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190144 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190244 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190275 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190302 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190326 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190351 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190381 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190408 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190434 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190459 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190482 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190508 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190536 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190563 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190657 4692 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190674 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190686 4692 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190702 4692 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190716 4692 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190728 4692 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190739 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185432 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.200942 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.201290 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.204047 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.204425 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.204434 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185741 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185902 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186064 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186089 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186385 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186516 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186554 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.186699 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187048 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187353 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187451 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187786 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187821 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188064 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.204724 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.187960 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188486 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.188836 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189017 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189155 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189762 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.189930 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190513 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190708 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.190895 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.191110 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.191147 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.191290 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.192089 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.192296 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.192720 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.192959 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.193061 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.193427 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.193669 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.193692 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.193708 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.195049 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.195335 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.195395 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.195400 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.195598 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.195641 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.196044 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.196084 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.196428 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.196464 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.196507 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.196560 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.197052 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.197487 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.197603 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.198053 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.198457 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.198615 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.199132 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.199209 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.205356 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.199576 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.200037 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.200407 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.204940 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.205872 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.206292 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.206543 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.206801 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.207280 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.207719 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.185715 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.207893 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.208107 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.208365 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.208426 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.208511 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.208653 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.208903 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.208944 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.209050 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.210220 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.210432 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.210491 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.210749 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.210825 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.210965 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.210936 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211199 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211331 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211488 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211589 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211731 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211946 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.212109 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:21:22.712076974 +0000 UTC m=+83.536812555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.212339 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.212378 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.213124 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.213492 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.213554 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.203533 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.216910 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211794 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.216935 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.214007 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.216957 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215006 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215122 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215318 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215476 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215725 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215785 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.211040 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215834 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215976 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215987 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215995 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.213613 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218062 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.216837 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.216854 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.216812 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.217066 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.217398 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.217543 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218356 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.217629 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215696 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215705 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.217985 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.215036 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218505 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218531 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218008 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218017 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218031 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218048 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218040 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218256 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218338 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218633 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.218774 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.219864 4692 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.219885 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.219943 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:22.719923934 +0000 UTC m=+83.544659515 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.220082 4692 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.220190 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.220580 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.200972 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.221064 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.221120 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.221149 4692 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.221199 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:22.721190852 +0000 UTC m=+83.545926433 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.221455 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.221752 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.222334 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.222424 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.222495 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.229029 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.229058 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.229148 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.229380 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.230561 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.230596 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.231213 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.231550 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.236017 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.237043 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.237104 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.237137 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.237213 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.238499 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.238544 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.238622 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.238652 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.238674 4692 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.238967 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:22.738933195 +0000 UTC m=+83.563668966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.239736 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.240687 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.242616 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.242812 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.242816 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.242858 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.243889 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.243919 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.243935 4692 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.244045 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:22.74400071 +0000 UTC m=+83.568736491 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.244100 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.244289 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.244613 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.244683 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.244735 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.245002 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.245200 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.245767 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.245971 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.246134 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.246484 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.246835 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.246873 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.248149 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.249285 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.259547 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.259592 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.259603 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.259620 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.259635 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.262782 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.264946 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.275031 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.279097 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291696 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291736 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291803 4692 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291815 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291825 4692 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291821 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291835 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291874 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291878 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291896 4692 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291912 4692 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291926 4692 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291937 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291947 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291956 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291969 4692 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291978 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.291989 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292003 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292017 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292029 4692 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292041 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292054 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292064 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292076 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292089 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292102 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292118 4692 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292130 4692 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292144 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292157 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292187 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292199 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292209 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292221 4692 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292233 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292244 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292258 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292270 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292281 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292292 4692 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292304 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292315 4692 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292328 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292340 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292352 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292363 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292377 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292389 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292400 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292410 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292422 4692 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292435 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292447 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292460 4692 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292471 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292480 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292490 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292500 4692 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292509 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292518 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292526 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292536 4692 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292545 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292554 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292564 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292573 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292582 4692 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292592 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292601 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292611 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292620 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292629 4692 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292638 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292648 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292657 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292666 4692 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292675 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292686 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292696 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292705 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292713 4692 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292722 4692 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292731 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292740 4692 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292749 4692 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292757 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292767 4692 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292777 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292807 4692 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292815 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292825 4692 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292834 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292842 4692 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292851 4692 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292860 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292869 4692 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292877 4692 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292887 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292896 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292906 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292917 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292926 4692 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292935 4692 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292945 4692 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292954 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292962 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292973 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292983 4692 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.292992 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293001 4692 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293011 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293020 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293030 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293038 4692 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293046 4692 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293054 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293064 4692 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293075 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293084 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293094 4692 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293103 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293112 4692 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293121 4692 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293132 4692 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293140 4692 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293149 4692 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293170 4692 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293179 4692 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293189 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293198 4692 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293207 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293217 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293227 4692 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293236 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293245 4692 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293254 4692 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293262 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293271 4692 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293280 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293290 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293301 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293310 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293319 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293328 4692 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293337 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293347 4692 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293355 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293364 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293373 4692 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293382 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293392 4692 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293401 4692 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293410 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293418 4692 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293427 4692 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293437 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293446 4692 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293455 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293658 4692 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293667 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293677 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293689 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293702 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293716 4692 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293729 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293742 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293754 4692 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293766 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293777 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293786 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293795 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293807 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293816 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293825 4692 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293835 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293845 4692 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293854 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293864 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293874 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293884 4692 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293895 4692 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293905 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293916 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293926 4692 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293935 4692 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293946 4692 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293955 4692 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293964 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293973 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.293982 4692 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.362356 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.362409 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.362429 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.362457 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.362478 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.363558 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.374087 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.382442 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 09 09:21:22 crc kubenswrapper[4692]: W0309 09:21:22.390735 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-63dacbc613b148f7fbd9e9de492ce70e2dd8ab26a6329486ef7640e8a88d37d7 WatchSource:0}: Error finding container 63dacbc613b148f7fbd9e9de492ce70e2dd8ab26a6329486ef7640e8a88d37d7: Status 404 returned error can't find the container with id 63dacbc613b148f7fbd9e9de492ce70e2dd8ab26a6329486ef7640e8a88d37d7 Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.394383 4692 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 09:21:22 crc kubenswrapper[4692]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 09 09:21:22 crc kubenswrapper[4692]: if [[ -f "/env/_master" ]]; then Mar 09 09:21:22 crc kubenswrapper[4692]: set -o allexport Mar 09 09:21:22 crc kubenswrapper[4692]: source "/env/_master" Mar 09 09:21:22 crc kubenswrapper[4692]: set +o allexport Mar 09 09:21:22 crc kubenswrapper[4692]: fi Mar 09 09:21:22 crc kubenswrapper[4692]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 09 09:21:22 crc kubenswrapper[4692]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 09 09:21:22 crc kubenswrapper[4692]: ho_enable="--enable-hybrid-overlay" Mar 09 09:21:22 crc kubenswrapper[4692]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 09 09:21:22 crc kubenswrapper[4692]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 09 09:21:22 crc kubenswrapper[4692]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 09 09:21:22 crc kubenswrapper[4692]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 09 09:21:22 crc kubenswrapper[4692]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 09 09:21:22 crc kubenswrapper[4692]: --webhook-host=127.0.0.1 \ Mar 09 09:21:22 crc kubenswrapper[4692]: --webhook-port=9743 \ Mar 09 09:21:22 crc kubenswrapper[4692]: ${ho_enable} \ Mar 09 09:21:22 crc kubenswrapper[4692]: --enable-interconnect \ Mar 09 09:21:22 crc kubenswrapper[4692]: --disable-approver \ Mar 09 09:21:22 crc kubenswrapper[4692]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 09 09:21:22 crc kubenswrapper[4692]: --wait-for-kubernetes-api=200s \ Mar 09 09:21:22 crc kubenswrapper[4692]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 09 09:21:22 crc kubenswrapper[4692]: --loglevel="${LOGLEVEL}" Mar 09 09:21:22 crc kubenswrapper[4692]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 09 09:21:22 crc kubenswrapper[4692]: > logger="UnhandledError" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.394971 4692 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.396203 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.398261 4692 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 09:21:22 crc kubenswrapper[4692]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 09 09:21:22 crc kubenswrapper[4692]: if [[ -f "/env/_master" ]]; then Mar 09 09:21:22 crc kubenswrapper[4692]: set -o allexport Mar 09 09:21:22 crc kubenswrapper[4692]: source "/env/_master" Mar 09 09:21:22 crc kubenswrapper[4692]: set +o allexport Mar 09 09:21:22 crc kubenswrapper[4692]: fi Mar 09 09:21:22 crc kubenswrapper[4692]: Mar 09 09:21:22 crc kubenswrapper[4692]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 09 09:21:22 crc kubenswrapper[4692]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 09 09:21:22 crc kubenswrapper[4692]: --disable-webhook \ Mar 09 09:21:22 crc kubenswrapper[4692]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 09 09:21:22 crc kubenswrapper[4692]: --loglevel="${LOGLEVEL}" Mar 09 09:21:22 crc kubenswrapper[4692]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 09 09:21:22 crc kubenswrapper[4692]: > logger="UnhandledError" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.399480 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 09 09:21:22 crc kubenswrapper[4692]: W0309 09:21:22.403665 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-f8d1a6faf2fa5df164cb4cb266597fcf77cf095a7b1b2ac2ba9cc991731512c6 WatchSource:0}: Error finding container f8d1a6faf2fa5df164cb4cb266597fcf77cf095a7b1b2ac2ba9cc991731512c6: Status 404 returned error can't find the container with id f8d1a6faf2fa5df164cb4cb266597fcf77cf095a7b1b2ac2ba9cc991731512c6 Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.408776 4692 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 09:21:22 crc kubenswrapper[4692]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 09 09:21:22 crc kubenswrapper[4692]: set -o allexport Mar 09 09:21:22 crc kubenswrapper[4692]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 09 09:21:22 crc kubenswrapper[4692]: source /etc/kubernetes/apiserver-url.env Mar 09 09:21:22 crc kubenswrapper[4692]: else Mar 09 09:21:22 crc kubenswrapper[4692]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 09 09:21:22 crc kubenswrapper[4692]: exit 1 Mar 09 09:21:22 crc kubenswrapper[4692]: fi Mar 09 09:21:22 crc kubenswrapper[4692]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 09 09:21:22 crc kubenswrapper[4692]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 09 09:21:22 crc kubenswrapper[4692]: > logger="UnhandledError" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.410519 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.465421 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.465465 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.465476 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.465493 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.465504 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.569640 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.569733 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.569756 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.569785 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.569806 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.673326 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.673400 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.673435 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.673466 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.673488 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.776589 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.776658 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.776678 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.776707 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.776728 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.797912 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.798078 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798094 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:21:23.798051275 +0000 UTC m=+84.622786896 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.798232 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.798291 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.798346 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798415 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798457 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798464 4692 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798485 4692 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798503 4692 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798576 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:23.798551741 +0000 UTC m=+84.623287352 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798611 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:23.798593672 +0000 UTC m=+84.623329293 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798642 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:23.798626463 +0000 UTC m=+84.623362084 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798687 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798727 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798747 4692 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: E0309 09:21:22.798822 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:23.798799048 +0000 UTC m=+84.623534659 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.879561 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.879648 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.879693 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.879733 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.879762 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.982521 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.982597 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.982623 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.982660 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:22 crc kubenswrapper[4692]: I0309 09:21:22.982687 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:22Z","lastTransitionTime":"2026-03-09T09:21:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.086101 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.086154 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.086195 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.086253 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.086264 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.188926 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.189002 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.189022 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.189050 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.189070 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.292513 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.292556 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.292565 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.292581 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.292591 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.387396 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f8d1a6faf2fa5df164cb4cb266597fcf77cf095a7b1b2ac2ba9cc991731512c6"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.389460 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"63dacbc613b148f7fbd9e9de492ce70e2dd8ab26a6329486ef7640e8a88d37d7"} Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.390476 4692 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 09:21:23 crc kubenswrapper[4692]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 09 09:21:23 crc kubenswrapper[4692]: set -o allexport Mar 09 09:21:23 crc kubenswrapper[4692]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 09 09:21:23 crc kubenswrapper[4692]: source /etc/kubernetes/apiserver-url.env Mar 09 09:21:23 crc kubenswrapper[4692]: else Mar 09 09:21:23 crc kubenswrapper[4692]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 09 09:21:23 crc kubenswrapper[4692]: exit 1 Mar 09 09:21:23 crc kubenswrapper[4692]: fi Mar 09 09:21:23 crc kubenswrapper[4692]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 09 09:21:23 crc kubenswrapper[4692]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 09 09:21:23 crc kubenswrapper[4692]: > logger="UnhandledError" Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.391574 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.391698 4692 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.392086 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"052bfc083b787ceeef0cc1ed6e9b56a9c4b74828cd3ad73c5d501e2cce258d22"} Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.393111 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.394421 4692 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 09:21:23 crc kubenswrapper[4692]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 09 09:21:23 crc kubenswrapper[4692]: if [[ -f "/env/_master" ]]; then Mar 09 09:21:23 crc kubenswrapper[4692]: set -o allexport Mar 09 09:21:23 crc kubenswrapper[4692]: source "/env/_master" Mar 09 09:21:23 crc kubenswrapper[4692]: set +o allexport Mar 09 09:21:23 crc kubenswrapper[4692]: fi Mar 09 09:21:23 crc kubenswrapper[4692]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 09 09:21:23 crc kubenswrapper[4692]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 09 09:21:23 crc kubenswrapper[4692]: ho_enable="--enable-hybrid-overlay" Mar 09 09:21:23 crc kubenswrapper[4692]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 09 09:21:23 crc kubenswrapper[4692]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 09 09:21:23 crc kubenswrapper[4692]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 09 09:21:23 crc kubenswrapper[4692]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 09 09:21:23 crc kubenswrapper[4692]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 09 09:21:23 crc kubenswrapper[4692]: --webhook-host=127.0.0.1 \ Mar 09 09:21:23 crc kubenswrapper[4692]: --webhook-port=9743 \ Mar 09 09:21:23 crc kubenswrapper[4692]: ${ho_enable} \ Mar 09 09:21:23 crc kubenswrapper[4692]: --enable-interconnect \ Mar 09 09:21:23 crc kubenswrapper[4692]: --disable-approver \ Mar 09 09:21:23 crc kubenswrapper[4692]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 09 09:21:23 crc kubenswrapper[4692]: --wait-for-kubernetes-api=200s \ Mar 09 09:21:23 crc kubenswrapper[4692]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 09 09:21:23 crc kubenswrapper[4692]: --loglevel="${LOGLEVEL}" Mar 09 09:21:23 crc kubenswrapper[4692]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 09 09:21:23 crc kubenswrapper[4692]: > logger="UnhandledError" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.394810 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.394838 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.394848 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.394865 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.394880 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.396994 4692 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 09 09:21:23 crc kubenswrapper[4692]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 09 09:21:23 crc kubenswrapper[4692]: if [[ -f "/env/_master" ]]; then Mar 09 09:21:23 crc kubenswrapper[4692]: set -o allexport Mar 09 09:21:23 crc kubenswrapper[4692]: source "/env/_master" Mar 09 09:21:23 crc kubenswrapper[4692]: set +o allexport Mar 09 09:21:23 crc kubenswrapper[4692]: fi Mar 09 09:21:23 crc kubenswrapper[4692]: Mar 09 09:21:23 crc kubenswrapper[4692]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 09 09:21:23 crc kubenswrapper[4692]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 09 09:21:23 crc kubenswrapper[4692]: --disable-webhook \ Mar 09 09:21:23 crc kubenswrapper[4692]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 09 09:21:23 crc kubenswrapper[4692]: --loglevel="${LOGLEVEL}" Mar 09 09:21:23 crc kubenswrapper[4692]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 09 09:21:23 crc kubenswrapper[4692]: > logger="UnhandledError" Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.398198 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.410427 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.426323 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.452861 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.474854 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.493802 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.497053 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.497093 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.497108 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.497126 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.497139 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.510275 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.521736 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.531303 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.542031 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.552257 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.562556 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.575741 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.600215 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.600277 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.600295 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.600321 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.600341 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.703312 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.703378 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.703402 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.703428 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.703446 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806061 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806279 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:21:25.80623291 +0000 UTC m=+86.630968531 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806342 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806404 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806440 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806450 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806468 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806481 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806432 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806560 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806613 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806620 4692 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806649 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806682 4692 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806686 4692 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806717 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:25.806698284 +0000 UTC m=+86.631434065 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.806655 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806751 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:25.806737035 +0000 UTC m=+86.631472876 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806753 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806776 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806777 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:25.806764796 +0000 UTC m=+86.631500407 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806787 4692 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:23 crc kubenswrapper[4692]: E0309 09:21:23.806844 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:25.806829428 +0000 UTC m=+86.631565009 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.910117 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.910191 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.910207 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.910228 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.910244 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:23Z","lastTransitionTime":"2026-03-09T09:21:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:23 crc kubenswrapper[4692]: I0309 09:21:23.992782 4692 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.012987 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.013044 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.013054 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.013073 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.013087 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.071655 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.071742 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.071795 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:24 crc kubenswrapper[4692]: E0309 09:21:24.071926 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:24 crc kubenswrapper[4692]: E0309 09:21:24.072047 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:24 crc kubenswrapper[4692]: E0309 09:21:24.072191 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.078708 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.079428 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.080590 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.081408 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.082286 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.082915 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.083657 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.084260 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.085062 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.085697 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.086401 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.087122 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.087815 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.089783 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.091060 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.092238 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.093644 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.094518 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.095964 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.097263 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.098307 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.100515 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.101129 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.101979 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.102459 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.103098 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.103885 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.104698 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.105427 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.105944 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.106426 4692 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.106523 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.107855 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.108408 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.108806 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.109973 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.112633 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.114089 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.115742 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.116775 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.116839 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.116860 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.116893 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.116915 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.117533 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.118065 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.119375 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.120042 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.121890 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.122958 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.125071 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.126223 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.127929 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.128443 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.129350 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.129862 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.130475 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.131500 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.131991 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.220201 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.220301 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.220327 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.220355 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.220377 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.324533 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.324605 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.324626 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.324656 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.324702 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.427750 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.427857 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.427876 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.427900 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.427919 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.531526 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.531608 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.531619 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.531637 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.531648 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.635409 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.635461 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.635475 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.635493 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.635513 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.738826 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.738873 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.738889 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.738909 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.738924 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.841960 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.842013 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.842027 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.842043 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.842055 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.945936 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.945999 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.946014 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.946036 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:24 crc kubenswrapper[4692]: I0309 09:21:24.946057 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:24Z","lastTransitionTime":"2026-03-09T09:21:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.049006 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.049044 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.049055 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.049070 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.049080 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.152026 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.152131 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.152152 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.152204 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.152230 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.254924 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.254973 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.254985 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.255001 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.255012 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.356792 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.356853 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.356863 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.356879 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.356889 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.459802 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.459871 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.459881 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.459893 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.459901 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.562865 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.562916 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.562926 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.562940 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.562949 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.665972 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.666046 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.666068 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.666097 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.666111 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.769542 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.769594 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.769610 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.769634 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.769652 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.824235 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.824353 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.824396 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824437 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:21:29.824409043 +0000 UTC m=+90.649144624 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.824472 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.824507 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824540 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824542 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824573 4692 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824581 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824603 4692 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824617 4692 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824619 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:29.824609819 +0000 UTC m=+90.649345400 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824559 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824660 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:29.82464944 +0000 UTC m=+90.649385021 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824664 4692 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824675 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:29.824667951 +0000 UTC m=+90.649403532 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:25 crc kubenswrapper[4692]: E0309 09:21:25.824728 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:29.824711052 +0000 UTC m=+90.649446643 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.873312 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.873366 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.873378 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.873406 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.873419 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.976754 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.976810 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.976844 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.976873 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:25 crc kubenswrapper[4692]: I0309 09:21:25.976891 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:25Z","lastTransitionTime":"2026-03-09T09:21:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.070686 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.070875 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.070967 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:26 crc kubenswrapper[4692]: E0309 09:21:26.071117 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:26 crc kubenswrapper[4692]: E0309 09:21:26.071296 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:26 crc kubenswrapper[4692]: E0309 09:21:26.071426 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.080868 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.080994 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.081024 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.081061 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.081087 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.184797 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.184847 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.184856 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.184875 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.184885 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.288051 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.288774 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.288877 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.288986 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.289087 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.393023 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.393095 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.393116 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.393146 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.393209 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.496653 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.496738 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.496761 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.496804 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.496825 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.600097 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.600144 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.600153 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.600190 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.600208 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.703291 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.703339 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.703348 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.703367 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.703378 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.805876 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.805944 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.805961 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.805990 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.806012 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.909107 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.909155 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.909192 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.909214 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:26 crc kubenswrapper[4692]: I0309 09:21:26.909228 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:26Z","lastTransitionTime":"2026-03-09T09:21:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.013100 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.013192 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.013208 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.013229 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.013244 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.082601 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.082730 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 09:21:27 crc kubenswrapper[4692]: E0309 09:21:27.082780 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.116668 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.116710 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.116720 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.116738 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.116751 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.219819 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.219874 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.219887 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.219905 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.219921 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.321801 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.321836 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.321845 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.321898 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.321912 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.402417 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:21:27 crc kubenswrapper[4692]: E0309 09:21:27.402555 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.424553 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.424606 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.424623 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.424648 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.424666 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.528217 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.528286 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.528310 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.528339 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.528362 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.630848 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.630905 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.630917 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.630945 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.630955 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.733078 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.733112 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.733120 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.733132 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.733143 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.836379 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.836428 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.836437 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.836453 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.836463 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.939521 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.939578 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.939590 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.939608 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.939628 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.962595 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.962643 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.962652 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.962665 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.962673 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: E0309 09:21:27.973265 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.978829 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.978869 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.978878 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.978895 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.978912 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:27 crc kubenswrapper[4692]: E0309 09:21:27.988362 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.992105 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.992137 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.992149 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.992187 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:27 crc kubenswrapper[4692]: I0309 09:21:27.992204 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:27Z","lastTransitionTime":"2026-03-09T09:21:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: E0309 09:21:28.005990 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.011151 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.011255 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.011275 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.011304 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.011323 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: E0309 09:21:28.023000 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.027264 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.027307 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.027316 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.027333 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.027344 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: E0309 09:21:28.036508 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:28 crc kubenswrapper[4692]: E0309 09:21:28.036629 4692 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.042434 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.042469 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.042480 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.042497 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.042508 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.070501 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:28 crc kubenswrapper[4692]: E0309 09:21:28.070643 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.070502 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:28 crc kubenswrapper[4692]: E0309 09:21:28.070732 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.070780 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:28 crc kubenswrapper[4692]: E0309 09:21:28.075916 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.145375 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.145424 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.145432 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.145449 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.145458 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.247917 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.247999 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.248010 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.248027 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.248039 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.351896 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.351947 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.351960 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.351977 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.351986 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.455906 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.455963 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.455980 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.456006 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.456025 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.559385 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.559465 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.559485 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.559512 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.559531 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.662770 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.662849 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.662863 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.662883 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.662895 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.766026 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.766525 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.766622 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.766719 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.766798 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.869778 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.869817 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.869828 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.869844 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.869857 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.976670 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.976733 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.976748 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.976770 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:28 crc kubenswrapper[4692]: I0309 09:21:28.976785 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:28Z","lastTransitionTime":"2026-03-09T09:21:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.079078 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.079130 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.079144 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.079155 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.079180 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.181078 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.181396 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.181505 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.181613 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.181702 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.284561 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.284611 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.284628 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.284644 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.284657 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.386376 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.386415 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.386427 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.386455 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.386469 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.489186 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.489226 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.489234 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.489248 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.489260 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.591887 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.591932 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.591941 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.591959 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.591969 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.694260 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.694303 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.694313 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.694326 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.694336 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.796509 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.796541 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.796549 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.796560 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.796568 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.865771 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.865843 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.865868 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.865897 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.865921 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.865992 4692 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.865992 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:21:37.865963639 +0000 UTC m=+98.690699240 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866039 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:37.866026721 +0000 UTC m=+98.690762302 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866100 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866133 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866153 4692 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866226 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:37.866210377 +0000 UTC m=+98.690945978 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866364 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866422 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866457 4692 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866362 4692 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866567 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:37.866524117 +0000 UTC m=+98.691259878 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:29 crc kubenswrapper[4692]: E0309 09:21:29.866834 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:37.866801925 +0000 UTC m=+98.691537676 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.898898 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.898998 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.899052 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.899090 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:29 crc kubenswrapper[4692]: I0309 09:21:29.899116 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:29Z","lastTransitionTime":"2026-03-09T09:21:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.001919 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.001964 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.001975 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.001993 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.002007 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.070690 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.070688 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:30 crc kubenswrapper[4692]: E0309 09:21:30.070846 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:30 crc kubenswrapper[4692]: E0309 09:21:30.070968 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.070707 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:30 crc kubenswrapper[4692]: E0309 09:21:30.071048 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.085277 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe078e2-7773-4e2e-b1d8-37078e375a7d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T09:21:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW0309 09:21:06.664316 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 09:21:06.664526 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 09:21:06.665180 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753571598/tls.crt::/tmp/serving-cert-3753571598/tls.key\\\\\\\"\\\\nI0309 09:21:06.974793 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 09:21:06.977569 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 09:21:06.977584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 09:21:06.977603 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 09:21:06.977609 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 09:21:06.984399 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 09:21:06.984421 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 09:21:06.984426 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 09:21:06.984430 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 09:21:06.984433 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 09:21:06.984437 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 09:21:06.984442 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 09:21:06.984455 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 09:21:06.985441 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T09:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T09:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T09:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T09:20:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.093492 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.103843 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.104454 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.104489 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.104504 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.104525 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.104540 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.113385 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.127211 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.134554 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.144978 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.206547 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.206591 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.206603 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.206620 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.206632 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.309276 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.309316 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.309325 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.309342 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.309351 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.410985 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.411028 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.411040 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.411057 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.411069 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.513879 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.513971 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.513989 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.514018 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.514037 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.617684 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.617747 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.617760 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.617779 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.617790 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.720316 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.720403 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.720421 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.720439 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.720451 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.823621 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.823683 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.823697 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.823719 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.823733 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.926520 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.926600 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.926623 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.926667 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:30 crc kubenswrapper[4692]: I0309 09:21:30.926688 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:30Z","lastTransitionTime":"2026-03-09T09:21:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.029969 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.030015 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.030026 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.030042 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.030052 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.133317 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.133376 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.133392 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.133415 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.133431 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.237319 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.237390 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.237409 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.237436 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.237458 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.341312 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.341358 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.341369 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.341386 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.341397 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.388573 4692 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.443915 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.443958 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.443971 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.443988 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.443999 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.545885 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.545948 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.545962 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.545978 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.545990 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.649104 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.649195 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.649216 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.649238 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.649256 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.751839 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.751912 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.751928 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.751948 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.751962 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.854213 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.854254 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.854300 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.854322 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.854335 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.957780 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.957823 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.957834 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.957851 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:31 crc kubenswrapper[4692]: I0309 09:21:31.957865 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:31Z","lastTransitionTime":"2026-03-09T09:21:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.059645 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.059692 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.059705 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.059721 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.059733 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.071355 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:32 crc kubenswrapper[4692]: E0309 09:21:32.071462 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.071737 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.071843 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:32 crc kubenswrapper[4692]: E0309 09:21:32.072278 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:32 crc kubenswrapper[4692]: E0309 09:21:32.072370 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.161844 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.161875 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.161885 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.161900 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.161911 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.264758 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.264802 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.264812 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.264828 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.264839 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.368610 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.368688 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.368700 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.368717 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.369098 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.471583 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.471630 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.471642 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.471657 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.471667 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.575139 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.575192 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.575204 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.575220 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.575231 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.677231 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.677281 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.677290 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.677306 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.677315 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.780050 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.780106 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.780124 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.780149 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.780193 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.884230 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.884310 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.884327 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.884382 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.884426 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.987546 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.987608 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.987619 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.987636 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:32 crc kubenswrapper[4692]: I0309 09:21:32.987648 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:32Z","lastTransitionTime":"2026-03-09T09:21:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.090728 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.090797 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.090818 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.090845 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.090864 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.194674 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.194731 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.194742 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.194762 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.194776 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.298047 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.298092 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.298103 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.298126 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.298139 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.400753 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.400804 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.400816 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.400834 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.400849 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.504081 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.504126 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.504136 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.504152 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.504180 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.607700 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.607756 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.607769 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.607788 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.607800 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.711245 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.711295 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.711304 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.711322 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.711334 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.815023 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.815094 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.815111 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.815137 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.815157 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.918741 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.918793 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.918803 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.918820 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:33 crc kubenswrapper[4692]: I0309 09:21:33.918830 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:33Z","lastTransitionTime":"2026-03-09T09:21:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.022410 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.022493 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.022517 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.022548 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.022575 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.071471 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.071492 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:34 crc kubenswrapper[4692]: E0309 09:21:34.071691 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.071618 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:34 crc kubenswrapper[4692]: E0309 09:21:34.071776 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:34 crc kubenswrapper[4692]: E0309 09:21:34.071833 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.126429 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.126520 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.126546 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.126580 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.126605 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.230466 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.230536 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.230549 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.230572 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.230587 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.333458 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.333520 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.333532 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.333553 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.333570 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.435437 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.435485 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.435495 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.435510 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.435520 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.538024 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.538085 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.538098 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.538120 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.538134 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.640666 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.640719 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.640731 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.640748 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.640760 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.743417 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.743466 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.743477 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.743493 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.743504 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.845681 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.845718 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.845728 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.845743 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.845753 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.948642 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.948715 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.948733 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.948758 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:34 crc kubenswrapper[4692]: I0309 09:21:34.948777 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:34Z","lastTransitionTime":"2026-03-09T09:21:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.051656 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.051744 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.051780 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.051821 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.051850 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.166279 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.166334 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.166347 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.166364 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.166377 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.268611 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.268659 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.268671 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.268688 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.268699 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.371041 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.371075 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.371085 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.371098 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.371106 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.587486 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.587522 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.587531 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.587544 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.587553 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.589855 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"60998552c63d1ad98d2aafd369a429a7664442a4fb5b8d506e2894631367841a"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.601706 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe078e2-7773-4e2e-b1d8-37078e375a7d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T09:21:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW0309 09:21:06.664316 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 09:21:06.664526 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 09:21:06.665180 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753571598/tls.crt::/tmp/serving-cert-3753571598/tls.key\\\\\\\"\\\\nI0309 09:21:06.974793 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 09:21:06.977569 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 09:21:06.977584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 09:21:06.977603 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 09:21:06.977609 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 09:21:06.984399 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 09:21:06.984421 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 09:21:06.984426 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 09:21:06.984430 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 09:21:06.984433 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 09:21:06.984437 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 09:21:06.984442 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 09:21:06.984455 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 09:21:06.985441 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T09:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T09:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T09:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T09:20:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.612495 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.622235 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.636368 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60998552c63d1ad98d2aafd369a429a7664442a4fb5b8d506e2894631367841a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:21:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.645825 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.653339 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.662799 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.689326 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.689383 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.689403 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.689425 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.689443 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.791640 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.791677 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.791688 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.791702 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.791711 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.894467 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.894501 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.894513 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.894527 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.894539 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.997711 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.997755 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.997767 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.997784 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:35 crc kubenswrapper[4692]: I0309 09:21:35.997796 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:35Z","lastTransitionTime":"2026-03-09T09:21:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.071399 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.071552 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:36 crc kubenswrapper[4692]: E0309 09:21:36.071665 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:36 crc kubenswrapper[4692]: E0309 09:21:36.071545 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.071420 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:36 crc kubenswrapper[4692]: E0309 09:21:36.071776 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.100373 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.100429 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.100444 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.100464 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.100477 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.202376 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.202422 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.202438 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.202462 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.202473 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.305301 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.305359 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.305376 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.305398 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.305415 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.409495 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.409549 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.409566 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.409594 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.409612 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.513249 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.513287 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.513301 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.513320 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.513334 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.615916 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.615980 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.615995 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.616020 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.616034 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.718507 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.718609 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.718636 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.718671 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.718798 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.821594 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.821633 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.821644 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.821660 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.821671 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.924354 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.924400 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.924412 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.924430 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.924442 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:36Z","lastTransitionTime":"2026-03-09T09:21:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:36 crc kubenswrapper[4692]: I0309 09:21:36.978436 4692 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.027342 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.027369 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.027378 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.027390 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.027398 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.130030 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.130498 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.130511 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.130531 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.130545 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.232409 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.232450 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.232460 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.232475 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.232486 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.334895 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.334939 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.334950 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.334968 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.334978 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.437683 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.437715 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.437724 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.437737 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.437745 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.540261 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.540317 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.540331 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.540347 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.540359 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.643968 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.644004 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.644013 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.644027 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.644037 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.746619 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.746680 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.746691 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.746708 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.746720 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.848845 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.848879 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.848889 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.848902 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.848913 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.906348 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.906476 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.906502 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.906526 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.906549 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906666 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:21:53.906621114 +0000 UTC m=+114.731356735 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906625 4692 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906728 4692 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906673 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906775 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:53.906756588 +0000 UTC m=+114.731492379 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906794 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:53.906785979 +0000 UTC m=+114.731521790 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906780 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906812 4692 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906839 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:53.90683221 +0000 UTC m=+114.731568011 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906675 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906865 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906875 4692 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:37 crc kubenswrapper[4692]: E0309 09:21:37.906904 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:53.906895672 +0000 UTC m=+114.731631463 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.951905 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.951950 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.951962 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.951980 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:37 crc kubenswrapper[4692]: I0309 09:21:37.951992 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:37Z","lastTransitionTime":"2026-03-09T09:21:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.055557 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.055618 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.055634 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.055661 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.055680 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.070874 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.070966 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.071075 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.071123 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.071660 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.071774 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.071873 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.072111 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.158446 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.158489 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.158503 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.158521 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.158534 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.260962 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.261021 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.261038 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.261058 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.261074 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.290876 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.290921 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.290931 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.290948 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.290960 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.299961 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.304193 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.304228 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.304240 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.304260 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.304273 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.313320 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.316809 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.316929 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.317007 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.317087 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.317194 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.326413 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.329911 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.329963 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.329973 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.329987 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.329998 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.337129 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.340899 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.340931 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.340940 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.340954 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.340964 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.348421 4692 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"09ec62b1-ccf9-4248-acae-6d6e8481578e\\\",\\\"systemUUID\\\":\\\"152c1cda-659c-435d-a04c-ef7c06479b86\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: E0309 09:21:38.348584 4692 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.363097 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.363148 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.363178 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.363199 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.363212 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.466195 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.466239 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.466249 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.466262 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.466271 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.568768 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.569034 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.569100 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.569171 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.569239 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.598613 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"887829cbf423ab21cead95b91df2d34a893dd9e11917e36e941fc7c08810f8d7"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.598877 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ae7933e99c97f814254c3007b9a0b16e329bf58d047f22cf43c29082dd6de894"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.610894 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:35Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60998552c63d1ad98d2aafd369a429a7664442a4fb5b8d506e2894631367841a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:21:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.618651 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.627596 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.635633 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.647208 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebe078e2-7773-4e2e-b1d8-37078e375a7d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-09T09:20:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-09T09:21:07Z\\\",\\\"message\\\":\\\"le observer\\\\nW0309 09:21:06.664316 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0309 09:21:06.664526 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0309 09:21:06.665180 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3753571598/tls.crt::/tmp/serving-cert-3753571598/tls.key\\\\\\\"\\\\nI0309 09:21:06.974793 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0309 09:21:06.977569 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0309 09:21:06.977584 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0309 09:21:06.977603 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0309 09:21:06.977609 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0309 09:21:06.984399 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0309 09:21:06.984421 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 09:21:06.984426 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0309 09:21:06.984430 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0309 09:21:06.984433 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0309 09:21:06.984437 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0309 09:21:06.984442 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0309 09:21:06.984455 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0309 09:21:06.985441 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-09T09:21:06Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:20:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-09T09:20:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-09T09:20:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-09T09:20:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.656129 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.672249 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.672316 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.672332 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.672358 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.672375 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.674766 4692 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:21:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://887829cbf423ab21cead95b91df2d34a893dd9e11917e36e941fc7c08810f8d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae7933e99c97f814254c3007b9a0b16e329bf58d047f22cf43c29082dd6de894\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-09T09:21:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.776028 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.776130 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.776220 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.776258 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.776282 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.879252 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.879325 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.879340 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.879366 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.879382 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.982315 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.982393 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.982415 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.982445 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:38 crc kubenswrapper[4692]: I0309 09:21:38.982471 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:38Z","lastTransitionTime":"2026-03-09T09:21:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.085520 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.085586 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.085604 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.085629 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.085648 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.192245 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.192414 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.192522 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.192566 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.192602 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.297061 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.297128 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.297147 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.297208 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.297228 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.399905 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.399963 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.399975 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.400263 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.400300 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.503570 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.503920 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.504068 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.504374 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.504603 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.606315 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.606462 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.606488 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.606510 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.606574 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.708990 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.709034 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.709050 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.709069 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.709084 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.811475 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.811509 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.811519 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.811534 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.811544 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.913951 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.914001 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.914018 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.914038 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:39 crc kubenswrapper[4692]: I0309 09:21:39.914063 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:39Z","lastTransitionTime":"2026-03-09T09:21:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.017444 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.017492 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.017507 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.017529 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.017544 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.070579 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.070550 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:40 crc kubenswrapper[4692]: E0309 09:21:40.070781 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.070866 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:40 crc kubenswrapper[4692]: E0309 09:21:40.071014 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:40 crc kubenswrapper[4692]: E0309 09:21:40.071117 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.119303 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.119348 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.119361 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.119376 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.119388 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.221176 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.221203 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.221212 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.221226 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.221239 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.323959 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.324003 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.324011 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.324024 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.324035 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.426259 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.426289 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.426298 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.426326 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.426337 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.529172 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.529208 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.529218 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.529233 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.529243 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.605414 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8f9a0cd5078d54cca1665ced624788e85531c220b063a3c1d4aca1d6d1d2d89a"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.631482 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.631545 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.631563 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.631590 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.631611 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.735988 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.736068 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.736094 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.736124 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.736146 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.839753 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.839819 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.839831 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.839849 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.839859 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.943559 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.943632 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.943645 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.943669 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:40 crc kubenswrapper[4692]: I0309 09:21:40.943684 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:40Z","lastTransitionTime":"2026-03-09T09:21:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.047104 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.047189 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.047202 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.047221 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.047232 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.153781 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.153845 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.153863 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.153892 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.153912 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.257082 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.257397 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.257475 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.257564 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.257642 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.360795 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.361046 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.361120 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.361257 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.361339 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.464537 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.464616 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.464635 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.464660 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.464680 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.568772 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.568845 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.568868 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.568899 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.568920 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.672822 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.672937 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.672962 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.672994 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.673017 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.777155 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.777262 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.777300 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.777330 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.777351 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.880963 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.881059 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.881078 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.881109 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.881128 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.985214 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.985294 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.985327 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.985363 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:41 crc kubenswrapper[4692]: I0309 09:21:41.985385 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:41Z","lastTransitionTime":"2026-03-09T09:21:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.071002 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:42 crc kubenswrapper[4692]: E0309 09:21:42.071350 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.071433 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.071479 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:42 crc kubenswrapper[4692]: E0309 09:21:42.071617 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:42 crc kubenswrapper[4692]: E0309 09:21:42.071920 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.088282 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.088352 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.088376 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.088413 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.088438 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.191837 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.191914 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.191933 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.191958 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.191976 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.294566 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.294607 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.294618 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.294642 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.294655 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.397309 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.397359 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.397373 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.397390 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.397403 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.499691 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.499776 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.499788 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.499806 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.499817 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.602133 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.602249 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.602282 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.602314 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.602335 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.705656 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.705727 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.705746 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.705778 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.705802 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.809974 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.810037 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.810055 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.810078 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.810098 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.912958 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.912991 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.912999 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.913011 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:42 crc kubenswrapper[4692]: I0309 09:21:42.913021 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:42Z","lastTransitionTime":"2026-03-09T09:21:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.015001 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.015042 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.015056 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.015101 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.015118 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.118048 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.118119 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.118143 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.118220 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.118249 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.221065 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.221105 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.221128 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.221142 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.221151 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.324029 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.324095 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.324113 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.324138 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.324196 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.427118 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.427203 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.427214 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.427228 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.427237 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.530356 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.530410 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.530422 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.530440 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.530452 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.633899 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.633965 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.633983 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.634011 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.634030 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.736259 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.736298 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.736311 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.736326 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.736335 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.839611 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.839655 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.839663 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.839685 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.839694 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.941857 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.941896 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.941905 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.941918 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:43 crc kubenswrapper[4692]: I0309 09:21:43.941928 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:43Z","lastTransitionTime":"2026-03-09T09:21:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.044716 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.044782 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.044797 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.044820 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.044838 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.070568 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:44 crc kubenswrapper[4692]: E0309 09:21:44.070756 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.070595 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.070863 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:44 crc kubenswrapper[4692]: E0309 09:21:44.070918 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:44 crc kubenswrapper[4692]: E0309 09:21:44.071037 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.147323 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.147371 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.147380 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.147410 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.147419 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.250384 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.250433 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.250447 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.250467 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.250482 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.352418 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.352465 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.352476 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.352520 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.352534 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.455272 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.455329 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.455340 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.455354 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.455364 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.557753 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.557840 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.557870 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.557900 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.557921 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.660556 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.660623 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.660705 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.660736 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.660759 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.763721 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.763824 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.763836 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.763851 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.763882 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.866984 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.867034 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.867045 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.867062 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.867074 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.969936 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.969993 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.970009 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.970037 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:44 crc kubenswrapper[4692]: I0309 09:21:44.970060 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:44Z","lastTransitionTime":"2026-03-09T09:21:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.072663 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.072702 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.072711 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.072740 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.072750 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.175542 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.175611 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.175628 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.175652 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.175670 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.279244 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.279289 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.279300 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.279315 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.279327 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.381766 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.382053 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.382127 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.382215 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.382296 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.485145 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.485214 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.485229 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.485250 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.485266 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.587585 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.587670 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.587696 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.587730 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.587753 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.690716 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.690763 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.690775 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.690794 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.690806 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.793429 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.793476 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.793492 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.793512 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.793527 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.895509 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.895551 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.895563 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.895595 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.895603 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.998149 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.998216 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.998225 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.998241 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:45 crc kubenswrapper[4692]: I0309 09:21:45.998267 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:45Z","lastTransitionTime":"2026-03-09T09:21:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.070764 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.070797 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.070884 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:46 crc kubenswrapper[4692]: E0309 09:21:46.071025 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:46 crc kubenswrapper[4692]: E0309 09:21:46.071084 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:46 crc kubenswrapper[4692]: E0309 09:21:46.071198 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.082347 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.100862 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.100903 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.100916 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.100954 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.100967 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.202857 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.202923 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.202936 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.202952 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.202963 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.305029 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.305059 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.305067 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.305080 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.305089 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.407819 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.407876 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.407912 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.407945 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.407973 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.510545 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.510589 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.510600 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.510616 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.510627 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.613174 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.613216 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.613233 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.613252 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.613264 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.716121 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.716217 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.716236 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.716255 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.716302 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.820001 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.820092 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.820128 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.820200 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.820226 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.923074 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.923102 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.923112 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.923124 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:46 crc kubenswrapper[4692]: I0309 09:21:46.923133 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:46Z","lastTransitionTime":"2026-03-09T09:21:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.026148 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.026222 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.026245 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.026266 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.026281 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.128802 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.128878 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.128903 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.128933 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.128955 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.231999 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.232045 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.232060 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.232079 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.232095 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.261259 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lmfk4"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.261727 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.264419 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.264426 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.264632 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.274291 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-f55bx"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.274982 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.277595 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.278560 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ndpxq"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.278779 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.278842 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.279229 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2cbtq"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.279433 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.279639 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.279740 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.282626 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.282703 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.283897 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.284240 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.284286 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.284507 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.285447 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.285720 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.291429 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-k72pq"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292171 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292541 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-system-cni-dir\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292588 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cnibin\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292620 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgqn7\" (UniqueName: \"kubernetes.io/projected/aa5521f7-9e8e-464b-81b7-4dc121a58f14-kube-api-access-bgqn7\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292645 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb18850a-c45f-438b-9854-5f8ced802c58-proxy-tls\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292669 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cni-binary-copy\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292810 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292874 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292900 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cb18850a-c45f-438b-9854-5f8ced802c58-rootfs\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.292922 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtkgn\" (UniqueName: \"kubernetes.io/projected/cb18850a-c45f-438b-9854-5f8ced802c58-kube-api-access-rtkgn\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.293032 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bslcl\" (UniqueName: \"kubernetes.io/projected/05606732-21a0-4b07-a3d7-f285c16344c1-kube-api-access-bslcl\") pod \"node-resolver-lmfk4\" (UID: \"05606732-21a0-4b07-a3d7-f285c16344c1\") " pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.293072 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-os-release\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.293094 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb18850a-c45f-438b-9854-5f8ced802c58-mcd-auth-proxy-config\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.293140 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/05606732-21a0-4b07-a3d7-f285c16344c1-hosts-file\") pod \"node-resolver-lmfk4\" (UID: \"05606732-21a0-4b07-a3d7-f285c16344c1\") " pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.294095 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.294541 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.294890 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.295136 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.295149 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.295667 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.297000 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.331943 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.331926065 podStartE2EDuration="1.331926065s" podCreationTimestamp="2026-03-09 09:21:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:47.313682427 +0000 UTC m=+108.138418068" watchObservedRunningTime="2026-03-09 09:21:47.331926065 +0000 UTC m=+108.156661666" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.333761 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.333807 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.333822 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.333841 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.333852 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394404 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394464 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-socket-dir-parent\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394486 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-daemon-config\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394508 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-script-lib\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394550 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtkgn\" (UniqueName: \"kubernetes.io/projected/cb18850a-c45f-438b-9854-5f8ced802c58-kube-api-access-rtkgn\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394570 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-netns\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394589 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-config\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394611 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394631 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-systemd-units\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394652 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-ovn-kubernetes\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394671 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-bin\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394688 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-netd\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394707 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-env-overrides\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394727 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-var-lib-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394746 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-etc-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394768 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394787 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-ovn\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394807 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-conf-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394826 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/008780c0-faad-4a8a-a6b4-4d092ff53b4a-cni-binary-copy\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394846 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-multus-certs\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394868 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smqfc\" (UniqueName: \"kubernetes.io/projected/008780c0-faad-4a8a-a6b4-4d092ff53b4a-kube-api-access-smqfc\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394889 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-log-socket\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394913 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bslcl\" (UniqueName: \"kubernetes.io/projected/05606732-21a0-4b07-a3d7-f285c16344c1-kube-api-access-bslcl\") pod \"node-resolver-lmfk4\" (UID: \"05606732-21a0-4b07-a3d7-f285c16344c1\") " pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394935 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb18850a-c45f-438b-9854-5f8ced802c58-mcd-auth-proxy-config\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.394982 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv54g\" (UniqueName: \"kubernetes.io/projected/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-kube-api-access-pv54g\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395010 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-k8s-cni-cncf-io\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395031 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgqn7\" (UniqueName: \"kubernetes.io/projected/aa5521f7-9e8e-464b-81b7-4dc121a58f14-kube-api-access-bgqn7\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395052 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-system-cni-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395073 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-system-cni-dir\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395094 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cni-binary-copy\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395115 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-systemd\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395147 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cb18850a-c45f-438b-9854-5f8ced802c58-rootfs\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395196 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-cni-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395218 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-cni-multus\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395250 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-cni-bin\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395270 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-netns\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395292 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395311 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-hostroot\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395330 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-kubelet\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395356 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-os-release\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395377 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-kubelet\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395444 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/05606732-21a0-4b07-a3d7-f285c16344c1-hosts-file\") pod \"node-resolver-lmfk4\" (UID: \"05606732-21a0-4b07-a3d7-f285c16344c1\") " pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395635 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-os-release\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395660 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-cnibin\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395706 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-node-log\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395733 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-system-cni-dir\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395737 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovn-node-metrics-cert\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395778 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb18850a-c45f-438b-9854-5f8ced802c58-proxy-tls\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395797 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-etc-kubernetes\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395815 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cnibin\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395831 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-os-release\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395845 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-slash\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395925 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cb18850a-c45f-438b-9854-5f8ced802c58-rootfs\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.395925 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cnibin\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.396072 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/05606732-21a0-4b07-a3d7-f285c16344c1-hosts-file\") pod \"node-resolver-lmfk4\" (UID: \"05606732-21a0-4b07-a3d7-f285c16344c1\") " pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.396416 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cni-binary-copy\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.396561 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb18850a-c45f-438b-9854-5f8ced802c58-mcd-auth-proxy-config\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.396563 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/aa5521f7-9e8e-464b-81b7-4dc121a58f14-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.396657 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/aa5521f7-9e8e-464b-81b7-4dc121a58f14-tuning-conf-dir\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.400976 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb18850a-c45f-438b-9854-5f8ced802c58-proxy-tls\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.412612 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bslcl\" (UniqueName: \"kubernetes.io/projected/05606732-21a0-4b07-a3d7-f285c16344c1-kube-api-access-bslcl\") pod \"node-resolver-lmfk4\" (UID: \"05606732-21a0-4b07-a3d7-f285c16344c1\") " pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.413897 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgqn7\" (UniqueName: \"kubernetes.io/projected/aa5521f7-9e8e-464b-81b7-4dc121a58f14-kube-api-access-bgqn7\") pod \"multus-additional-cni-plugins-f55bx\" (UID: \"aa5521f7-9e8e-464b-81b7-4dc121a58f14\") " pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.415285 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtkgn\" (UniqueName: \"kubernetes.io/projected/cb18850a-c45f-438b-9854-5f8ced802c58-kube-api-access-rtkgn\") pod \"machine-config-daemon-ndpxq\" (UID: \"cb18850a-c45f-438b-9854-5f8ced802c58\") " pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.436558 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.436606 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.436616 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.436630 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.436642 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.440685 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-4qtml"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.441043 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.442960 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.443050 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.443403 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.445846 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496382 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-system-cni-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496423 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-systemd\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496469 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-cni-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496498 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-cni-multus\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496518 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-cni-bin\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496534 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-netns\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496551 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-hostroot\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496555 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-system-cni-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496570 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-kubelet\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496609 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-kubelet\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496625 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496655 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-cni-bin\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496671 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-kubelet\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496682 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-systemd\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496691 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-cni-multus\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496719 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-var-lib-kubelet\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496719 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-cni-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496748 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-hostroot\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496745 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-netns\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496765 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496745 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-cnibin\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496808 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-node-log\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496848 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovn-node-metrics-cert\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.496969 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-etc-kubernetes\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497026 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-etc-kubernetes\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497063 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-os-release\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497151 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-slash\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497186 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-cnibin\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497217 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-node-log\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497241 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-slash\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497242 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-os-release\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497356 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-socket-dir-parent\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497380 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-daemon-config\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497396 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-script-lib\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497432 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-netns\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497610 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-netns\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497646 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-config\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497679 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-systemd-units\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497703 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-ovn-kubernetes\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497724 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-bin\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497744 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-netd\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497763 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-env-overrides\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497766 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-systemd-units\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497783 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-conf-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497800 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-socket-dir-parent\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497805 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-var-lib-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497811 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-netd\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497835 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-var-lib-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497837 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-etc-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497862 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-conf-dir\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497871 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497879 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-ovn-kubernetes\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497897 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-ovn\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497903 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-etc-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497924 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/008780c0-faad-4a8a-a6b4-4d092ff53b4a-cni-binary-copy\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497933 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-openvswitch\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497948 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-multus-certs\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497952 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-ovn\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497973 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smqfc\" (UniqueName: \"kubernetes.io/projected/008780c0-faad-4a8a-a6b4-4d092ff53b4a-kube-api-access-smqfc\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.497994 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-log-socket\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498003 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-multus-certs\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498040 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-k8s-cni-cncf-io\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498015 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/008780c0-faad-4a8a-a6b4-4d092ff53b4a-host-run-k8s-cni-cncf-io\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498123 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/008780c0-faad-4a8a-a6b4-4d092ff53b4a-multus-daemon-config\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498203 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-log-socket\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498218 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv54g\" (UniqueName: \"kubernetes.io/projected/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-kube-api-access-pv54g\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498435 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-env-overrides\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498457 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-config\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498629 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/008780c0-faad-4a8a-a6b4-4d092ff53b4a-cni-binary-copy\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.498692 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-bin\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.499349 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-script-lib\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.499747 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovn-node-metrics-cert\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.514255 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smqfc\" (UniqueName: \"kubernetes.io/projected/008780c0-faad-4a8a-a6b4-4d092ff53b4a-kube-api-access-smqfc\") pod \"multus-2cbtq\" (UID: \"008780c0-faad-4a8a-a6b4-4d092ff53b4a\") " pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.515689 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv54g\" (UniqueName: \"kubernetes.io/projected/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-kube-api-access-pv54g\") pod \"ovnkube-node-k72pq\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.538940 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.538978 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.538987 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.539000 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.539008 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.565736 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.566149 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.567647 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.568279 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.584227 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9dn4z"] Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.584859 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:47 crc kubenswrapper[4692]: E0309 09:21:47.584928 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9dn4z" podUID="68ff6f00-0ac5-41d9-b762-933944bdb507" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.593549 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lmfk4" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598723 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d9a7db8-2ab0-417a-9199-77bb992c3490-host\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598755 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598771 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8d9a7db8-2ab0-417a-9199-77bb992c3490-serviceca\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598795 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598820 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598878 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2qmz\" (UniqueName: \"kubernetes.io/projected/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-kube-api-access-b2qmz\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598893 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfmfz\" (UniqueName: \"kubernetes.io/projected/68ff6f00-0ac5-41d9-b762-933944bdb507-kube-api-access-mfmfz\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.598918 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.599012 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6kzv\" (UniqueName: \"kubernetes.io/projected/8d9a7db8-2ab0-417a-9199-77bb992c3490-kube-api-access-r6kzv\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.607034 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-f55bx" Mar 09 09:21:47 crc kubenswrapper[4692]: W0309 09:21:47.619560 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa5521f7_9e8e_464b_81b7_4dc121a58f14.slice/crio-4f9c1cde77735680c7a572d9b453ef2d105e60bb47f7271d467e673b8f83ffce WatchSource:0}: Error finding container 4f9c1cde77735680c7a572d9b453ef2d105e60bb47f7271d467e673b8f83ffce: Status 404 returned error can't find the container with id 4f9c1cde77735680c7a572d9b453ef2d105e60bb47f7271d467e673b8f83ffce Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.623079 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.626151 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lmfk4" event={"ID":"05606732-21a0-4b07-a3d7-f285c16344c1","Type":"ContainerStarted","Data":"9fd6a6a37673df5b551361316b6b3abd234b3cf1abd5cee5be0c9ea515692710"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.627057 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerStarted","Data":"4f9c1cde77735680c7a572d9b453ef2d105e60bb47f7271d467e673b8f83ffce"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.639651 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2cbtq" Mar 09 09:21:47 crc kubenswrapper[4692]: W0309 09:21:47.641918 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb18850a_c45f_438b_9854_5f8ced802c58.slice/crio-d8a1e66d5644949c01d4e68fe47b8403dfe73dba896cd3766026413bb5c7ae15 WatchSource:0}: Error finding container d8a1e66d5644949c01d4e68fe47b8403dfe73dba896cd3766026413bb5c7ae15: Status 404 returned error can't find the container with id d8a1e66d5644949c01d4e68fe47b8403dfe73dba896cd3766026413bb5c7ae15 Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.642013 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.642047 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.642059 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.642075 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.642086 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.648791 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:47 crc kubenswrapper[4692]: W0309 09:21:47.652873 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod008780c0_faad_4a8a_a6b4_4d092ff53b4a.slice/crio-86002bec8fff957f8bc0e0ae4304cf7bf45310fe576f01de9625fe7057add688 WatchSource:0}: Error finding container 86002bec8fff957f8bc0e0ae4304cf7bf45310fe576f01de9625fe7057add688: Status 404 returned error can't find the container with id 86002bec8fff957f8bc0e0ae4304cf7bf45310fe576f01de9625fe7057add688 Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.700656 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.700990 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2qmz\" (UniqueName: \"kubernetes.io/projected/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-kube-api-access-b2qmz\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701009 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfmfz\" (UniqueName: \"kubernetes.io/projected/68ff6f00-0ac5-41d9-b762-933944bdb507-kube-api-access-mfmfz\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701042 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701061 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6kzv\" (UniqueName: \"kubernetes.io/projected/8d9a7db8-2ab0-417a-9199-77bb992c3490-kube-api-access-r6kzv\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701085 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701111 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d9a7db8-2ab0-417a-9199-77bb992c3490-host\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701137 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8d9a7db8-2ab0-417a-9199-77bb992c3490-serviceca\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701201 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.701836 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: E0309 09:21:47.701995 4692 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:47 crc kubenswrapper[4692]: E0309 09:21:47.702038 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs podName:68ff6f00-0ac5-41d9-b762-933944bdb507 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:48.202025404 +0000 UTC m=+109.026760985 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs") pod "network-metrics-daemon-9dn4z" (UID: "68ff6f00-0ac5-41d9-b762-933944bdb507") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.702199 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8d9a7db8-2ab0-417a-9199-77bb992c3490-host\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.703037 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8d9a7db8-2ab0-417a-9199-77bb992c3490-serviceca\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.706225 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.708420 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.717090 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2qmz\" (UniqueName: \"kubernetes.io/projected/4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2-kube-api-access-b2qmz\") pod \"ovnkube-control-plane-749d76644c-p4jks\" (UID: \"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.718304 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6kzv\" (UniqueName: \"kubernetes.io/projected/8d9a7db8-2ab0-417a-9199-77bb992c3490-kube-api-access-r6kzv\") pod \"node-ca-4qtml\" (UID: \"8d9a7db8-2ab0-417a-9199-77bb992c3490\") " pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.719354 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfmfz\" (UniqueName: \"kubernetes.io/projected/68ff6f00-0ac5-41d9-b762-933944bdb507-kube-api-access-mfmfz\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.748586 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.748623 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.748636 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.748651 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.748662 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.752400 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4qtml" Mar 09 09:21:47 crc kubenswrapper[4692]: W0309 09:21:47.771521 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d9a7db8_2ab0_417a_9199_77bb992c3490.slice/crio-dd4f5129d33cddee260b82d66ee5bea7d854fe26a53610ad2b05831150f731b9 WatchSource:0}: Error finding container dd4f5129d33cddee260b82d66ee5bea7d854fe26a53610ad2b05831150f731b9: Status 404 returned error can't find the container with id dd4f5129d33cddee260b82d66ee5bea7d854fe26a53610ad2b05831150f731b9 Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.852360 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.852389 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.852398 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.852413 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.852422 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.876524 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" Mar 09 09:21:47 crc kubenswrapper[4692]: W0309 09:21:47.887325 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c4972ef_a3e9_46f6_b30a_3c3e5c8daae2.slice/crio-09459ffadca8f93490d807c726caba34584b41ce340812a0d9b6a10cb9748ddb WatchSource:0}: Error finding container 09459ffadca8f93490d807c726caba34584b41ce340812a0d9b6a10cb9748ddb: Status 404 returned error can't find the container with id 09459ffadca8f93490d807c726caba34584b41ce340812a0d9b6a10cb9748ddb Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.955800 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.955846 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.955856 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.955872 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:47 crc kubenswrapper[4692]: I0309 09:21:47.955881 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:47Z","lastTransitionTime":"2026-03-09T09:21:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.057905 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.057934 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.057944 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.057957 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.057966 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:48Z","lastTransitionTime":"2026-03-09T09:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.071303 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.071333 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.071321 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:48 crc kubenswrapper[4692]: E0309 09:21:48.071432 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:48 crc kubenswrapper[4692]: E0309 09:21:48.071544 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:48 crc kubenswrapper[4692]: E0309 09:21:48.071633 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.160570 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.160612 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.160623 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.160639 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.160651 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:48Z","lastTransitionTime":"2026-03-09T09:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.207591 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:48 crc kubenswrapper[4692]: E0309 09:21:48.207831 4692 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:48 crc kubenswrapper[4692]: E0309 09:21:48.207921 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs podName:68ff6f00-0ac5-41d9-b762-933944bdb507 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:49.207901676 +0000 UTC m=+110.032637457 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs") pod "network-metrics-daemon-9dn4z" (UID: "68ff6f00-0ac5-41d9-b762-933944bdb507") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.263415 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.263456 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.263470 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.263486 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.263496 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:48Z","lastTransitionTime":"2026-03-09T09:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.366145 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.366200 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.366212 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.366225 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.366235 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:48Z","lastTransitionTime":"2026-03-09T09:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.469315 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.469368 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.469382 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.469402 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.469414 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:48Z","lastTransitionTime":"2026-03-09T09:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.572443 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.572489 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.572501 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.572518 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.572529 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:48Z","lastTransitionTime":"2026-03-09T09:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.635295 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" event={"ID":"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2","Type":"ContainerStarted","Data":"ddf33219e8d13c2b4c6624ca22ce43ecd98c8c2941e1fc53b495a8767a1d2147"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.635348 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" event={"ID":"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2","Type":"ContainerStarted","Data":"cb0df8c63487e8aec1d407c0c569b748c1dc246787603cec15bdca2480a77646"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.635365 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" event={"ID":"4c4972ef-a3e9-46f6-b30a-3c3e5c8daae2","Type":"ContainerStarted","Data":"09459ffadca8f93490d807c726caba34584b41ce340812a0d9b6a10cb9748ddb"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.637382 4692 generic.go:334] "Generic (PLEG): container finished" podID="aa5521f7-9e8e-464b-81b7-4dc121a58f14" containerID="0b947ebc11b35cc934e3c948400c04df961c32fb3e4c990dcae3d00407ef029c" exitCode=0 Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.637449 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerDied","Data":"0b947ebc11b35cc934e3c948400c04df961c32fb3e4c990dcae3d00407ef029c"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.639276 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955" exitCode=0 Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.639383 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.639415 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"2f2e2f093b91bba6196566d45d5dd232e73fcd4194220a9c0396b2bd8ce9431d"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.641436 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2cbtq" event={"ID":"008780c0-faad-4a8a-a6b4-4d092ff53b4a","Type":"ContainerStarted","Data":"c6ec87e2b80080204e209614ace6f0ad6b7cf9165ba1078af6ffd2d5214a6d6f"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.641471 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2cbtq" event={"ID":"008780c0-faad-4a8a-a6b4-4d092ff53b4a","Type":"ContainerStarted","Data":"86002bec8fff957f8bc0e0ae4304cf7bf45310fe576f01de9625fe7057add688"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.642821 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4qtml" event={"ID":"8d9a7db8-2ab0-417a-9199-77bb992c3490","Type":"ContainerStarted","Data":"b5f751919893379e5399a0475754cd91e178e9ff79de56de1b6b62cec1a894ee"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.642872 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4qtml" event={"ID":"8d9a7db8-2ab0-417a-9199-77bb992c3490","Type":"ContainerStarted","Data":"dd4f5129d33cddee260b82d66ee5bea7d854fe26a53610ad2b05831150f731b9"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.645525 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"0b79a0c10d18c69391ab8c838aa6aa2cb1f48a806ba89c71b7945882e0aa87be"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.645592 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"b30ec1f14935ee72de8d696ab21516444d1dbed8e188660550c2b615537f1de4"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.645620 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"d8a1e66d5644949c01d4e68fe47b8403dfe73dba896cd3766026413bb5c7ae15"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.646700 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lmfk4" event={"ID":"05606732-21a0-4b07-a3d7-f285c16344c1","Type":"ContainerStarted","Data":"3ed4ed4022c40810fab21bfe174ab2451348096f2eee35065c75b4fa56a5ece1"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.666042 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p4jks" podStartSLOduration=43.666019737 podStartE2EDuration="43.666019737s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:48.664724217 +0000 UTC m=+109.489459798" watchObservedRunningTime="2026-03-09 09:21:48.666019737 +0000 UTC m=+109.490755358" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.669884 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.669916 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.669924 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.669937 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.669946 4692 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-09T09:21:48Z","lastTransitionTime":"2026-03-09T09:21:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.741551 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lmfk4" podStartSLOduration=44.741527826 podStartE2EDuration="44.741527826s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:48.696882781 +0000 UTC m=+109.521618372" watchObservedRunningTime="2026-03-09 09:21:48.741527826 +0000 UTC m=+109.566263417" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.747023 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-4qtml" podStartSLOduration=44.747004124 podStartE2EDuration="44.747004124s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:48.742069713 +0000 UTC m=+109.566805304" watchObservedRunningTime="2026-03-09 09:21:48.747004124 +0000 UTC m=+109.571739715" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.747410 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv"] Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.747805 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.750899 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.751114 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.751409 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.751614 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.756131 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podStartSLOduration=44.756112543 podStartE2EDuration="44.756112543s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:48.75536643 +0000 UTC m=+109.580102031" watchObservedRunningTime="2026-03-09 09:21:48.756112543 +0000 UTC m=+109.580848134" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.790025 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2cbtq" podStartSLOduration=43.790009209 podStartE2EDuration="43.790009209s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:48.789472963 +0000 UTC m=+109.614208544" watchObservedRunningTime="2026-03-09 09:21:48.790009209 +0000 UTC m=+109.614744780" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.822043 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11a9e8d-60a2-4733-a3b1-9875d5b93846-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.822091 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c11a9e8d-60a2-4733-a3b1-9875d5b93846-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.822125 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11a9e8d-60a2-4733-a3b1-9875d5b93846-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.822149 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c11a9e8d-60a2-4733-a3b1-9875d5b93846-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.822274 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c11a9e8d-60a2-4733-a3b1-9875d5b93846-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.923129 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c11a9e8d-60a2-4733-a3b1-9875d5b93846-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.923194 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11a9e8d-60a2-4733-a3b1-9875d5b93846-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.923216 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c11a9e8d-60a2-4733-a3b1-9875d5b93846-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.923241 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11a9e8d-60a2-4733-a3b1-9875d5b93846-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.923258 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c11a9e8d-60a2-4733-a3b1-9875d5b93846-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.923301 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c11a9e8d-60a2-4733-a3b1-9875d5b93846-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.923314 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c11a9e8d-60a2-4733-a3b1-9875d5b93846-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.924138 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c11a9e8d-60a2-4733-a3b1-9875d5b93846-service-ca\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.934789 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c11a9e8d-60a2-4733-a3b1-9875d5b93846-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:48 crc kubenswrapper[4692]: I0309 09:21:48.937848 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c11a9e8d-60a2-4733-a3b1-9875d5b93846-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-qjfzv\" (UID: \"c11a9e8d-60a2-4733-a3b1-9875d5b93846\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.071363 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:49 crc kubenswrapper[4692]: E0309 09:21:49.071530 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9dn4z" podUID="68ff6f00-0ac5-41d9-b762-933944bdb507" Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.077698 4692 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.085399 4692 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.110261 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" Mar 09 09:21:49 crc kubenswrapper[4692]: W0309 09:21:49.123822 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc11a9e8d_60a2_4733_a3b1_9875d5b93846.slice/crio-2b461549900034280174a779b00c02842dd453cdf695b7fbbcdd2b876e04c355 WatchSource:0}: Error finding container 2b461549900034280174a779b00c02842dd453cdf695b7fbbcdd2b876e04c355: Status 404 returned error can't find the container with id 2b461549900034280174a779b00c02842dd453cdf695b7fbbcdd2b876e04c355 Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.225619 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:49 crc kubenswrapper[4692]: E0309 09:21:49.225918 4692 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:49 crc kubenswrapper[4692]: E0309 09:21:49.226697 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs podName:68ff6f00-0ac5-41d9-b762-933944bdb507 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:51.226679524 +0000 UTC m=+112.051415105 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs") pod "network-metrics-daemon-9dn4z" (UID: "68ff6f00-0ac5-41d9-b762-933944bdb507") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.651259 4692 generic.go:334] "Generic (PLEG): container finished" podID="aa5521f7-9e8e-464b-81b7-4dc121a58f14" containerID="e79db7d0a5d99575f298a9ab6d3bfd2e0877407ac704665a37fde50104a68830" exitCode=0 Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.651330 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerDied","Data":"e79db7d0a5d99575f298a9ab6d3bfd2e0877407ac704665a37fde50104a68830"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.653470 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" event={"ID":"c11a9e8d-60a2-4733-a3b1-9875d5b93846","Type":"ContainerStarted","Data":"7c25fdfe0c3d1ebea44eb95805daad33cd30395030c2b43d8e1e0753e89e60c7"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.653511 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" event={"ID":"c11a9e8d-60a2-4733-a3b1-9875d5b93846","Type":"ContainerStarted","Data":"2b461549900034280174a779b00c02842dd453cdf695b7fbbcdd2b876e04c355"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.658460 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.658506 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.658519 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.658530 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.658541 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} Mar 09 09:21:49 crc kubenswrapper[4692]: I0309 09:21:49.658551 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} Mar 09 09:21:50 crc kubenswrapper[4692]: I0309 09:21:50.071312 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:50 crc kubenswrapper[4692]: I0309 09:21:50.071435 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:50 crc kubenswrapper[4692]: I0309 09:21:50.072379 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:50 crc kubenswrapper[4692]: E0309 09:21:50.072365 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:50 crc kubenswrapper[4692]: E0309 09:21:50.072634 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:50 crc kubenswrapper[4692]: E0309 09:21:50.072715 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:50 crc kubenswrapper[4692]: I0309 09:21:50.663796 4692 generic.go:334] "Generic (PLEG): container finished" podID="aa5521f7-9e8e-464b-81b7-4dc121a58f14" containerID="69043be37990aef5fe59c8624aeacdcbcbefb4c98de7a7b531c6665ab8237a04" exitCode=0 Mar 09 09:21:50 crc kubenswrapper[4692]: I0309 09:21:50.663880 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerDied","Data":"69043be37990aef5fe59c8624aeacdcbcbefb4c98de7a7b531c6665ab8237a04"} Mar 09 09:21:50 crc kubenswrapper[4692]: I0309 09:21:50.688370 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-qjfzv" podStartSLOduration=45.688352457 podStartE2EDuration="45.688352457s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:49.68758959 +0000 UTC m=+110.512325171" watchObservedRunningTime="2026-03-09 09:21:50.688352457 +0000 UTC m=+111.513088038" Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.071184 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:51 crc kubenswrapper[4692]: E0309 09:21:51.071463 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9dn4z" podUID="68ff6f00-0ac5-41d9-b762-933944bdb507" Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.072253 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.247762 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:51 crc kubenswrapper[4692]: E0309 09:21:51.248004 4692 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:51 crc kubenswrapper[4692]: E0309 09:21:51.248373 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs podName:68ff6f00-0ac5-41d9-b762-933944bdb507 nodeName:}" failed. No retries permitted until 2026-03-09 09:21:55.248351335 +0000 UTC m=+116.073086926 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs") pod "network-metrics-daemon-9dn4z" (UID: "68ff6f00-0ac5-41d9-b762-933944bdb507") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.671152 4692 generic.go:334] "Generic (PLEG): container finished" podID="aa5521f7-9e8e-464b-81b7-4dc121a58f14" containerID="07dff4ce99baedbcc1285ac4404335b3d2e0472a58cd2c996f742cb930f36e0d" exitCode=0 Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.671221 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerDied","Data":"07dff4ce99baedbcc1285ac4404335b3d2e0472a58cd2c996f742cb930f36e0d"} Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.676068 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.677722 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245"} Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.678048 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:21:51 crc kubenswrapper[4692]: I0309 09:21:51.723670 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=24.723653081 podStartE2EDuration="24.723653081s" podCreationTimestamp="2026-03-09 09:21:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:51.723094844 +0000 UTC m=+112.547830425" watchObservedRunningTime="2026-03-09 09:21:51.723653081 +0000 UTC m=+112.548388662" Mar 09 09:21:52 crc kubenswrapper[4692]: I0309 09:21:52.071462 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:52 crc kubenswrapper[4692]: E0309 09:21:52.071593 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:52 crc kubenswrapper[4692]: I0309 09:21:52.071984 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:52 crc kubenswrapper[4692]: E0309 09:21:52.072060 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:52 crc kubenswrapper[4692]: I0309 09:21:52.074651 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:52 crc kubenswrapper[4692]: E0309 09:21:52.074729 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:52 crc kubenswrapper[4692]: I0309 09:21:52.682840 4692 generic.go:334] "Generic (PLEG): container finished" podID="aa5521f7-9e8e-464b-81b7-4dc121a58f14" containerID="d881542d3041f74434ca390b6da3e8a29a7a27f862e313c724b138fd65cde109" exitCode=0 Mar 09 09:21:52 crc kubenswrapper[4692]: I0309 09:21:52.682898 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerDied","Data":"d881542d3041f74434ca390b6da3e8a29a7a27f862e313c724b138fd65cde109"} Mar 09 09:21:52 crc kubenswrapper[4692]: I0309 09:21:52.696804 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.071538 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.071749 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9dn4z" podUID="68ff6f00-0ac5-41d9-b762-933944bdb507" Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.702025 4692 generic.go:334] "Generic (PLEG): container finished" podID="aa5521f7-9e8e-464b-81b7-4dc121a58f14" containerID="42ae335a768935de38b5f289be3468091c3a76ba7ecc5833be309bb3ee635540" exitCode=0 Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.702127 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerDied","Data":"42ae335a768935de38b5f289be3468091c3a76ba7ecc5833be309bb3ee635540"} Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.975395 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.975531 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.975550 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:25.975525122 +0000 UTC m=+146.800260723 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.975579 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.975620 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.975642 4692 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: I0309 09:21:53.975656 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.975680 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:25.975672076 +0000 UTC m=+146.800407657 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.975767 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.975782 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.975795 4692 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.975828 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:25.975817001 +0000 UTC m=+146.800552582 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.976128 4692 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.976169 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:25.976154271 +0000 UTC m=+146.800889852 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.976244 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.976259 4692 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.976270 4692 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:53 crc kubenswrapper[4692]: E0309 09:21:53.976298 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:25.976288785 +0000 UTC m=+146.801024366 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.071439 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.071459 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.071515 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:54 crc kubenswrapper[4692]: E0309 09:21:54.071569 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:54 crc kubenswrapper[4692]: E0309 09:21:54.071659 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:54 crc kubenswrapper[4692]: E0309 09:21:54.071745 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.711427 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-f55bx" event={"ID":"aa5521f7-9e8e-464b-81b7-4dc121a58f14","Type":"ContainerStarted","Data":"7a5ce297801167b23960cfeedc46292b316a10d3ded71e642e7e1cce9d9d7e5b"} Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.717590 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerStarted","Data":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.717979 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.753692 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.769118 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-f55bx" podStartSLOduration=49.769095333 podStartE2EDuration="49.769095333s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:54.735293089 +0000 UTC m=+115.560028700" watchObservedRunningTime="2026-03-09 09:21:54.769095333 +0000 UTC m=+115.593830934" Mar 09 09:21:54 crc kubenswrapper[4692]: I0309 09:21:54.769328 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podStartSLOduration=49.769318209 podStartE2EDuration="49.769318209s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:21:54.768710461 +0000 UTC m=+115.593446062" watchObservedRunningTime="2026-03-09 09:21:54.769318209 +0000 UTC m=+115.594053810" Mar 09 09:21:55 crc kubenswrapper[4692]: I0309 09:21:55.071289 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:55 crc kubenswrapper[4692]: E0309 09:21:55.071454 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9dn4z" podUID="68ff6f00-0ac5-41d9-b762-933944bdb507" Mar 09 09:21:55 crc kubenswrapper[4692]: I0309 09:21:55.291730 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:55 crc kubenswrapper[4692]: E0309 09:21:55.291898 4692 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:55 crc kubenswrapper[4692]: E0309 09:21:55.291969 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs podName:68ff6f00-0ac5-41d9-b762-933944bdb507 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.291951563 +0000 UTC m=+124.116687154 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs") pod "network-metrics-daemon-9dn4z" (UID: "68ff6f00-0ac5-41d9-b762-933944bdb507") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 09 09:21:55 crc kubenswrapper[4692]: I0309 09:21:55.721976 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:55 crc kubenswrapper[4692]: I0309 09:21:55.722530 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:55 crc kubenswrapper[4692]: I0309 09:21:55.757900 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:21:56 crc kubenswrapper[4692]: I0309 09:21:56.071266 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:56 crc kubenswrapper[4692]: I0309 09:21:56.071296 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:56 crc kubenswrapper[4692]: I0309 09:21:56.071317 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:56 crc kubenswrapper[4692]: E0309 09:21:56.071395 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:56 crc kubenswrapper[4692]: E0309 09:21:56.071771 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:56 crc kubenswrapper[4692]: E0309 09:21:56.071830 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:56 crc kubenswrapper[4692]: I0309 09:21:56.547047 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9dn4z"] Mar 09 09:21:56 crc kubenswrapper[4692]: I0309 09:21:56.547214 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:56 crc kubenswrapper[4692]: E0309 09:21:56.547314 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9dn4z" podUID="68ff6f00-0ac5-41d9-b762-933944bdb507" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.071204 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.071230 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.071240 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:21:58 crc kubenswrapper[4692]: E0309 09:21:58.071309 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.071371 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:21:58 crc kubenswrapper[4692]: E0309 09:21:58.071440 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 09 09:21:58 crc kubenswrapper[4692]: E0309 09:21:58.071490 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 09 09:21:58 crc kubenswrapper[4692]: E0309 09:21:58.071643 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9dn4z" podUID="68ff6f00-0ac5-41d9-b762-933944bdb507" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.651145 4692 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.651355 4692 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.689269 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nxm8t"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.689594 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.689793 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.689852 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.692369 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.692604 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.692831 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.693757 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.694046 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.694133 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.694249 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.695589 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.697301 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.699354 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.699548 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.699783 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.700150 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.701189 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7gw55"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.701892 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.702428 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.705454 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-4sdrj"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.716748 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.717033 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.717198 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zd4xm"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.717283 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.717887 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.717928 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9vg42"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.718392 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.718810 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.718864 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.719007 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.722657 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pbl6c"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.723091 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qpk8p"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.724450 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.724558 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.724593 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.724784 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.725080 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.725201 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.725772 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.725799 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.725932 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726043 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726147 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726324 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726470 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726658 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726673 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726749 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726813 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726916 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.726984 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727097 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-serving-cert\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727120 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727125 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-encryption-config\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727148 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727196 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727220 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-config\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727251 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mwq5\" (UniqueName: \"kubernetes.io/projected/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-kube-api-access-4mwq5\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727289 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nncr4\" (UniqueName: \"kubernetes.io/projected/33a84922-5581-4432-b6d9-d5a9edd7defe-kube-api-access-nncr4\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727309 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-serving-cert\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727329 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45f39c-f9bd-4115-83f4-e866533d1090-audit-dir\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727362 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727388 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a84922-5581-4432-b6d9-d5a9edd7defe-config\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727412 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gsf9\" (UniqueName: \"kubernetes.io/projected/ca45f39c-f9bd-4115-83f4-e866533d1090-kube-api-access-8gsf9\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727433 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/33a84922-5581-4432-b6d9-d5a9edd7defe-images\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727455 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-config\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727476 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-serving-cert\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727495 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-audit-policies\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727514 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-client-ca\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727547 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-client-ca\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727568 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22vg9\" (UniqueName: \"kubernetes.io/projected/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-kube-api-access-22vg9\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727596 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/33a84922-5581-4432-b6d9-d5a9edd7defe-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727617 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-etcd-client\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.727257 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.729590 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.729713 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.733066 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.734887 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.735291 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.735460 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.735607 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.735814 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.736014 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rfkdj"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.736442 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.736686 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.736924 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.737019 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.737250 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.737727 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.737750 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.737738 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.737951 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.738050 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.738126 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-snq8b"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.738149 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.738567 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-rddqb"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.738801 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.738831 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.738970 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxwb2"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.739109 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-snq8b" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.739394 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.739438 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.739477 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.739476 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.740289 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.740455 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.740583 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.741281 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.741570 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.741789 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.742059 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.741589 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.740168 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.753943 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.763766 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.763835 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.764227 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.764429 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.765241 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.765412 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.765554 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.765655 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.765789 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.765883 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766167 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766247 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766329 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766428 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766494 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766524 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766563 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766643 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766778 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766833 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766950 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766787 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.767049 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.767127 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.767198 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.767270 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.767320 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.767402 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.767418 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.766170 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.768251 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.768662 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.781799 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.781906 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.782032 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.782154 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.782335 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.782512 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.783453 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.784164 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.784880 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.785466 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.791117 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.795322 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.798797 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.798878 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.799967 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.800484 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.801708 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.801934 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.802440 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.802805 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.803433 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.806548 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7gw55"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.807819 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.807934 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.817117 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828604 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828731 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nncr4\" (UniqueName: \"kubernetes.io/projected/33a84922-5581-4432-b6d9-d5a9edd7defe-kube-api-access-nncr4\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828804 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-serving-cert\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828841 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45f39c-f9bd-4115-83f4-e866533d1090-audit-dir\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828869 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph2ts\" (UniqueName: \"kubernetes.io/projected/d59db20e-6720-4cc9-bfd8-b3477c7298cf-kube-api-access-ph2ts\") pod \"dns-operator-744455d44c-rfkdj\" (UID: \"d59db20e-6720-4cc9-bfd8-b3477c7298cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828898 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97ncr\" (UniqueName: \"kubernetes.io/projected/553737d8-1213-41ef-a0ee-d1f85d749634-kube-api-access-97ncr\") pod \"cluster-samples-operator-665b6dd947-tr7zk\" (UID: \"553737d8-1213-41ef-a0ee-d1f85d749634\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828930 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc559\" (UniqueName: \"kubernetes.io/projected/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-kube-api-access-qc559\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.828975 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvvzb\" (UniqueName: \"kubernetes.io/projected/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-kube-api-access-lvvzb\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829005 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829033 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-audit-policies\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829071 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8385d993-0021-4066-a9c6-57bf7ee4be33-metrics-tls\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829100 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829143 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829193 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-config\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829228 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj8pl\" (UniqueName: \"kubernetes.io/projected/f73e489b-1a01-4608-a493-e0ce7dd2c2c5-kube-api-access-wj8pl\") pod \"migrator-59844c95c7-ftqvs\" (UID: \"f73e489b-1a01-4608-a493-e0ce7dd2c2c5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829255 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829285 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-oauth-config\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829317 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a84922-5581-4432-b6d9-d5a9edd7defe-config\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829351 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmvx4\" (UniqueName: \"kubernetes.io/projected/d7e9d340-f169-49ff-82ae-21089bdb28f6-kube-api-access-lmvx4\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829380 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gsf9\" (UniqueName: \"kubernetes.io/projected/ca45f39c-f9bd-4115-83f4-e866533d1090-kube-api-access-8gsf9\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829412 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829445 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/33a84922-5581-4432-b6d9-d5a9edd7defe-images\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829526 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-config\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829577 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829610 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829636 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d59db20e-6720-4cc9-bfd8-b3477c7298cf-metrics-tls\") pod \"dns-operator-744455d44c-rfkdj\" (UID: \"d59db20e-6720-4cc9-bfd8-b3477c7298cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829671 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-serving-cert\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829694 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-audit-policies\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829717 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8385d993-0021-4066-a9c6-57bf7ee4be33-trusted-ca\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829719 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45f39c-f9bd-4115-83f4-e866533d1090-audit-dir\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.829741 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5537a456-f771-4a55-b811-f20b50cd6446-audit-dir\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.830464 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33a84922-5581-4432-b6d9-d5a9edd7defe-config\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.830953 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.831172 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.831223 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-client-ca\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.831245 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.831303 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-client-ca\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.831681 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/33a84922-5581-4432-b6d9-d5a9edd7defe-images\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.831788 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-audit-policies\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.833534 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e9d340-f169-49ff-82ae-21089bdb28f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.833683 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22vg9\" (UniqueName: \"kubernetes.io/projected/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-kube-api-access-22vg9\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.833807 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.833977 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-serving-cert\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.834276 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/33a84922-5581-4432-b6d9-d5a9edd7defe-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.834478 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-trusted-ca-bundle\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.834499 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.834797 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-client-ca\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.835118 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.835467 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-client-ca\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.835752 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrpmz\" (UniqueName: \"kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-kube-api-access-lrpmz\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.835827 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-etcd-client\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836263 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e9d340-f169-49ff-82ae-21089bdb28f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836281 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836412 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfwlq\" (UniqueName: \"kubernetes.io/projected/5537a456-f771-4a55-b811-f20b50cd6446-kube-api-access-kfwlq\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836489 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-serving-cert\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836360 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836667 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-encryption-config\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836772 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836887 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.836991 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837104 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-service-ca\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837230 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837299 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/33a84922-5581-4432-b6d9-d5a9edd7defe-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837403 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837504 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837607 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-oauth-serving-cert\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837712 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-config\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.837826 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mwq5\" (UniqueName: \"kubernetes.io/projected/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-kube-api-access-4mwq5\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.838121 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.838244 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.838335 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/553737d8-1213-41ef-a0ee-d1f85d749634-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tr7zk\" (UID: \"553737d8-1213-41ef-a0ee-d1f85d749634\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.838133 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.838929 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45f39c-f9bd-4115-83f4-e866533d1090-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.839001 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-etcd-client\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.839152 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-config\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.841644 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.842027 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-serving-cert\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.843309 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-config\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.843475 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45f39c-f9bd-4115-83f4-e866533d1090-encryption-config\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.844951 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.845560 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.846080 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdvhn"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.846644 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vdjpl"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.847403 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.847982 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.848863 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.849279 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.849578 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.850061 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.850557 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.851126 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-serving-cert\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.851243 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.851797 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.855598 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.854394 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-serving-cert\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.860677 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.862048 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.862304 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.863674 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.864353 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.868665 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.873602 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jsgfm"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.873839 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.874865 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nxm8t"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.875158 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.875521 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.875725 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rxzs8"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.875983 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.876252 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.876940 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.877052 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jdvnz"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.877213 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.879156 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.881483 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-d5xx5"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.881871 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.884509 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.886116 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qpk8p"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.886420 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9vg42"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.887351 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rddqb"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.888189 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rfkdj"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.889119 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.890617 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxwb2"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.892054 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.892228 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vdjpl"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.892250 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.896839 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.897810 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zd4xm"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.899489 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.903208 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.904498 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-snq8b"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.905083 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.906756 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.908328 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.909212 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.910407 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.911137 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.911871 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdvhn"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.912463 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.913562 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.914664 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vbp6d"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.915458 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vbp6d" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.915682 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.917890 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.918833 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.920167 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vbp6d"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.920634 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jsgfm"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.921680 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jdvnz"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.922930 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-4sdrj"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.923940 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pbl6c"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.924888 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.925846 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.926900 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rxzs8"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.928296 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.930029 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.931025 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.931718 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-c9rx4"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.932479 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.934030 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fct6h"] Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.934738 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.938918 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8385d993-0021-4066-a9c6-57bf7ee4be33-trusted-ca\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.938942 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5537a456-f771-4a55-b811-f20b50cd6446-audit-dir\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.938963 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.938982 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939010 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e9d340-f169-49ff-82ae-21089bdb28f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939032 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939048 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5537a456-f771-4a55-b811-f20b50cd6446-audit-dir\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939089 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-serving-cert\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939112 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-trusted-ca-bundle\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939129 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrpmz\" (UniqueName: \"kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-kube-api-access-lrpmz\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939188 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e9d340-f169-49ff-82ae-21089bdb28f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939207 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfwlq\" (UniqueName: \"kubernetes.io/projected/5537a456-f771-4a55-b811-f20b50cd6446-kube-api-access-kfwlq\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939232 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939250 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939264 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-service-ca\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939286 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939306 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939324 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-oauth-serving-cert\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939345 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939366 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/553737d8-1213-41ef-a0ee-d1f85d749634-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tr7zk\" (UID: \"553737d8-1213-41ef-a0ee-d1f85d749634\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.939392 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940447 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97ncr\" (UniqueName: \"kubernetes.io/projected/553737d8-1213-41ef-a0ee-d1f85d749634-kube-api-access-97ncr\") pod \"cluster-samples-operator-665b6dd947-tr7zk\" (UID: \"553737d8-1213-41ef-a0ee-d1f85d749634\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940494 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc559\" (UniqueName: \"kubernetes.io/projected/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-kube-api-access-qc559\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940527 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph2ts\" (UniqueName: \"kubernetes.io/projected/d59db20e-6720-4cc9-bfd8-b3477c7298cf-kube-api-access-ph2ts\") pod \"dns-operator-744455d44c-rfkdj\" (UID: \"d59db20e-6720-4cc9-bfd8-b3477c7298cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940551 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvvzb\" (UniqueName: \"kubernetes.io/projected/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-kube-api-access-lvvzb\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940598 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-audit-policies\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940633 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8385d993-0021-4066-a9c6-57bf7ee4be33-metrics-tls\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940657 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.940676 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.941544 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-config\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.943843 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.944730 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.944814 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.945024 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-audit-policies\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.945282 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e9d340-f169-49ff-82ae-21089bdb28f6-config\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946064 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-config\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946132 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-oauth-config\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946187 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj8pl\" (UniqueName: \"kubernetes.io/projected/f73e489b-1a01-4608-a493-e0ce7dd2c2c5-kube-api-access-wj8pl\") pod \"migrator-59844c95c7-ftqvs\" (UID: \"f73e489b-1a01-4608-a493-e0ce7dd2c2c5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946220 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946259 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmvx4\" (UniqueName: \"kubernetes.io/projected/d7e9d340-f169-49ff-82ae-21089bdb28f6-kube-api-access-lmvx4\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946306 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946337 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946363 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.946395 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d59db20e-6720-4cc9-bfd8-b3477c7298cf-metrics-tls\") pod \"dns-operator-744455d44c-rfkdj\" (UID: \"d59db20e-6720-4cc9-bfd8-b3477c7298cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.947465 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e9d340-f169-49ff-82ae-21089bdb28f6-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.947693 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.948043 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/553737d8-1213-41ef-a0ee-d1f85d749634-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-tr7zk\" (UID: \"553737d8-1213-41ef-a0ee-d1f85d749634\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.948385 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.948445 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.948457 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.949444 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.951619 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.951771 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.952301 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.953594 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.953617 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d59db20e-6720-4cc9-bfd8-b3477c7298cf-metrics-tls\") pod \"dns-operator-744455d44c-rfkdj\" (UID: \"d59db20e-6720-4cc9-bfd8-b3477c7298cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.953756 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.957335 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.972835 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.977735 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-serving-cert\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:58 crc kubenswrapper[4692]: I0309 09:21:58.991532 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.000151 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-console-oauth-config\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.012462 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.031671 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.035500 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-service-ca\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.057769 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.060455 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-trusted-ca-bundle\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.071589 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.076635 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-oauth-serving-cert\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.091921 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.112085 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.132246 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.152885 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.171952 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.178022 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8385d993-0021-4066-a9c6-57bf7ee4be33-metrics-tls\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.197204 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.201198 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8385d993-0021-4066-a9c6-57bf7ee4be33-trusted-ca\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.212329 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.231871 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.251859 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.272811 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.312001 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.333055 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.352386 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.372516 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.425834 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gsf9\" (UniqueName: \"kubernetes.io/projected/ca45f39c-f9bd-4115-83f4-e866533d1090-kube-api-access-8gsf9\") pod \"apiserver-7bbb656c7d-bhnwz\" (UID: \"ca45f39c-f9bd-4115-83f4-e866533d1090\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.445488 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nncr4\" (UniqueName: \"kubernetes.io/projected/33a84922-5581-4432-b6d9-d5a9edd7defe-kube-api-access-nncr4\") pod \"machine-api-operator-5694c8668f-7gw55\" (UID: \"33a84922-5581-4432-b6d9-d5a9edd7defe\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.466537 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22vg9\" (UniqueName: \"kubernetes.io/projected/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-kube-api-access-22vg9\") pod \"controller-manager-879f6c89f-nxm8t\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.471563 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.492086 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.511794 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.531695 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.551663 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.571653 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.592436 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.618953 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.625100 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mwq5\" (UniqueName: \"kubernetes.io/projected/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-kube-api-access-4mwq5\") pod \"route-controller-manager-6576b87f9c-q286z\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.625864 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.632426 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.652260 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.652489 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.662564 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.671962 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.703019 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.714042 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.732838 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.754556 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.772648 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.796701 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.811283 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.832554 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.852002 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.870437 4692 request.go:700] Waited for 1.019896132s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.872150 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.893328 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.911733 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.931358 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.951781 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.971492 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 09 09:21:59 crc kubenswrapper[4692]: I0309 09:21:59.992692 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.013662 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.027901 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nxm8t"] Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.029300 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z"] Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.033823 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.052382 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.071575 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.071615 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.071612 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.071766 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.073090 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.084545 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7gw55"] Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.087366 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz"] Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.092418 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.111728 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.123719 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550802-w82g5"] Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.125246 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550802-w82g5" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.131252 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550802-w82g5"] Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.131872 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.154847 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.172530 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.191521 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.211960 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.232550 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.251868 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.272425 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.292093 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.312232 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.331923 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.352437 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.372063 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.391470 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.411769 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.432310 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.451560 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.471942 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.492107 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.512272 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.532049 4692 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.553112 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.573114 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.592985 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.612602 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.632434 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.653292 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.672410 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.692756 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.711440 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.732735 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.743262 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" event={"ID":"1e53fbe9-5027-4dae-8f6f-f2eda177f94b","Type":"ContainerStarted","Data":"02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.743324 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" event={"ID":"1e53fbe9-5027-4dae-8f6f-f2eda177f94b","Type":"ContainerStarted","Data":"04e00af1f96d5952625bde42114285e09711fbc65d39383ef21157e084f1acf7"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.744231 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.745994 4692 generic.go:334] "Generic (PLEG): container finished" podID="ca45f39c-f9bd-4115-83f4-e866533d1090" containerID="3a8544a9ddfd231e1963f9b140409821babdf06662cd8cff7b64b574af872624" exitCode=0 Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.746078 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" event={"ID":"ca45f39c-f9bd-4115-83f4-e866533d1090","Type":"ContainerDied","Data":"3a8544a9ddfd231e1963f9b140409821babdf06662cd8cff7b64b574af872624"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.746099 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" event={"ID":"ca45f39c-f9bd-4115-83f4-e866533d1090","Type":"ContainerStarted","Data":"305c3dd30433e2b325a4bf158a9560c3f27987f260e983a8cf588a5ca336e97c"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.746274 4692 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nxm8t container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.746465 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" podUID="1e53fbe9-5027-4dae-8f6f-f2eda177f94b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.747825 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" event={"ID":"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb","Type":"ContainerStarted","Data":"c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.747883 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" event={"ID":"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb","Type":"ContainerStarted","Data":"3a9d0ad7db8c22a125ddd7d08097b231dc348773f75fb465e29000d2ae09f95b"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.748071 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.750367 4692 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-q286z container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.750447 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" podUID="4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.752742 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.757412 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" event={"ID":"33a84922-5581-4432-b6d9-d5a9edd7defe","Type":"ContainerStarted","Data":"6cdeaf23b6cdee9e5d5a76c108d6618ec6819fe070d5e250a2a166ddff1e7ffc"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.757459 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" event={"ID":"33a84922-5581-4432-b6d9-d5a9edd7defe","Type":"ContainerStarted","Data":"a884aed69d40140790b0f899b2604e981f4ee36aeae6da8ba02281b9043c831b"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.757470 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" event={"ID":"33a84922-5581-4432-b6d9-d5a9edd7defe","Type":"ContainerStarted","Data":"f5a54725cb80f55c2f2233b20b7b65f81ea115eb370c38b5f36ab18813e8e304"} Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.772099 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.792911 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.813365 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.832915 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.870472 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.870879 4692 request.go:700] Waited for 1.931483362s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.912589 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:22:00 crc kubenswrapper[4692]: I0309 09:22:00.988576 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvvzb\" (UniqueName: \"kubernetes.io/projected/0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758-kube-api-access-lvvzb\") pod \"cluster-image-registry-operator-dc59b4c8b-8hxvb\" (UID: \"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.091876 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.113104 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.132049 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.152220 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.169962 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946e899e-f89e-4f66-80a2-1f3effc06b7f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170009 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-audit\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170065 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-certificates\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170086 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946e899e-f89e-4f66-80a2-1f3effc06b7f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170192 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-config\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170219 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddda8795-d770-4f73-a1c2-c988092f7fa4-config\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170267 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cbcb6e-701d-45c4-8293-945716329cef-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170292 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-ca\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170336 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm5sx\" (UniqueName: \"kubernetes.io/projected/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-kube-api-access-bm5sx\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170360 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/146e15b3-9b76-4a85-acb5-85530b4fbdc2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170425 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-serving-cert\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170504 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-trusted-ca\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170567 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-etcd-serving-ca\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170621 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-serving-cert\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170646 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t88ck\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-kube-api-access-t88ck\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170712 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv2qt\" (UniqueName: \"kubernetes.io/projected/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-kube-api-access-sv2qt\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170763 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170785 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d24t6\" (UniqueName: \"kubernetes.io/projected/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-kube-api-access-d24t6\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170848 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-encryption-config\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170875 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-config\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170931 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-trusted-ca\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.170985 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-config\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171011 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-service-ca\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171063 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-serving-cert\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171131 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ddda8795-d770-4f73-a1c2-c988092f7fa4-machine-approver-tls\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171236 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/146e15b3-9b76-4a85-acb5-85530b4fbdc2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171263 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdwrf\" (UniqueName: \"kubernetes.io/projected/4607c813-5202-460e-b583-a03dc13eee01-kube-api-access-tdwrf\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171337 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ddda8795-d770-4f73-a1c2-c988092f7fa4-auth-proxy-config\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171418 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fcz2\" (UniqueName: \"kubernetes.io/projected/c279ea73-49f4-441d-9a36-bedd8388437e-kube-api-access-4fcz2\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171454 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/64cbcb6e-701d-45c4-8293-945716329cef-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171471 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-serving-cert\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171490 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171513 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-etcd-client\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171531 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nzq9\" (UniqueName: \"kubernetes.io/projected/ddda8795-d770-4f73-a1c2-c988092f7fa4-kube-api-access-8nzq9\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171588 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c279ea73-49f4-441d-9a36-bedd8388437e-audit-dir\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171620 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-client\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171661 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4607c813-5202-460e-b583-a03dc13eee01-serving-cert\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171707 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mwqz\" (UniqueName: \"kubernetes.io/projected/aa908134-de3c-43e0-ace1-94fc548861c8-kube-api-access-5mwqz\") pod \"downloads-7954f5f757-snq8b\" (UID: \"aa908134-de3c-43e0-ace1-94fc548861c8\") " pod="openshift-console/downloads-7954f5f757-snq8b" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171735 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-tls\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171754 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-config\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171776 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-service-ca-bundle\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171807 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-image-import-ca\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171836 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171855 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cbcb6e-701d-45c4-8293-945716329cef-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171889 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c279ea73-49f4-441d-9a36-bedd8388437e-node-pullsecrets\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171913 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vtnd\" (UniqueName: \"kubernetes.io/projected/146e15b3-9b76-4a85-acb5-85530b4fbdc2-kube-api-access-6vtnd\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171931 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-bound-sa-token\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.171948 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.172463 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:01.672449902 +0000 UTC m=+122.497185733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.175367 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.191520 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.211757 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.232401 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.251537 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272279 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.272424 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:01.772399898 +0000 UTC m=+122.597135499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272473 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-config\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272495 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-service-ca\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272513 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-serving-cert\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272528 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ddda8795-d770-4f73-a1c2-c988092f7fa4-machine-approver-tls\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272568 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ddda8795-d770-4f73-a1c2-c988092f7fa4-auth-proxy-config\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272601 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/64cbcb6e-701d-45c4-8293-945716329cef-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272617 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-serving-cert\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272637 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-plugins-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272654 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272664 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nzq9\" (UniqueName: \"kubernetes.io/projected/ddda8795-d770-4f73-a1c2-c988092f7fa4-kube-api-access-8nzq9\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272770 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/d4063de5-2cdb-4d86-a74f-c98af0aae38d-ready\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272862 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4607c813-5202-460e-b583-a03dc13eee01-serving-cert\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272887 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7880dcf7-b6e3-4620-ba8b-013f207b02e0-cert\") pod \"ingress-canary-jsgfm\" (UID: \"7880dcf7-b6e3-4620-ba8b-013f207b02e0\") " pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272908 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf7zs\" (UniqueName: \"kubernetes.io/projected/03b56c7f-0613-419a-8a32-fe678a4a3e93-kube-api-access-gf7zs\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272933 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-tls\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272962 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mwqz\" (UniqueName: \"kubernetes.io/projected/aa908134-de3c-43e0-ace1-94fc548861c8-kube-api-access-5mwqz\") pod \"downloads-7954f5f757-snq8b\" (UID: \"aa908134-de3c-43e0-ace1-94fc548861c8\") " pod="openshift-console/downloads-7954f5f757-snq8b" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.272985 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/543c31b5-3ba9-4e04-aeaf-3ba7315da541-proxy-tls\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273041 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/700952b9-c6cb-403f-855b-b3cbd855e845-tmpfs\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273063 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f93524c6-53a2-4df0-873b-241f5ff69a9d-signing-key\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273082 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273109 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273125 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-socket-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273163 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c279ea73-49f4-441d-9a36-bedd8388437e-node-pullsecrets\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273196 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93ccefab-c47d-4725-a9ff-55cc50f678f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vdjpl\" (UID: \"93ccefab-c47d-4725-a9ff-55cc50f678f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273255 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273269 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-srv-cert\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273307 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df306321-4c89-4909-834c-48ae131ac196-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273324 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbw5p\" (UniqueName: \"kubernetes.io/projected/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-kube-api-access-sbw5p\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273340 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-metrics-tls\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273371 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-config\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273392 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c86z\" (UniqueName: \"kubernetes.io/projected/543c31b5-3ba9-4e04-aeaf-3ba7315da541-kube-api-access-8c86z\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273408 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f93524c6-53a2-4df0-873b-241f5ff69a9d-signing-cabundle\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273424 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb95cf2-78dc-43a2-acce-73f8be8e9f03-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273500 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzd4h\" (UniqueName: \"kubernetes.io/projected/cb17c405-b001-4e0a-8721-776ce83db499-kube-api-access-qzd4h\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273518 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273565 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm5sx\" (UniqueName: \"kubernetes.io/projected/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-kube-api-access-bm5sx\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273589 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/146e15b3-9b76-4a85-acb5-85530b4fbdc2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273607 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-serving-cert\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273623 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273638 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb17c405-b001-4e0a-8721-776ce83db499-service-ca-bundle\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273652 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-mountpoint-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273666 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df306321-4c89-4909-834c-48ae131ac196-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273682 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-srv-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273695 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-csi-data-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273739 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f530da-08e1-42ee-a635-ca648ee4f5b3-config\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273774 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-encryption-config\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.273796 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-config\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.274891 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-trusted-ca\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.274922 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c9ac14-6e46-4243-a490-6467522ae8e4-config\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.274946 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/700952b9-c6cb-403f-855b-b3cbd855e845-apiservice-cert\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.275240 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvrr9\" (UniqueName: \"kubernetes.io/projected/f93524c6-53a2-4df0-873b-241f5ff69a9d-kube-api-access-nvrr9\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.275275 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/974389aa-40cb-4191-a85b-d14e1aa175af-proxy-tls\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.275318 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bng8j\" (UID: \"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.275345 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsqmf\" (UniqueName: \"kubernetes.io/projected/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-kube-api-access-wsqmf\") pod \"control-plane-machine-set-operator-78cbb6b69f-bng8j\" (UID: \"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.275373 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/146e15b3-9b76-4a85-acb5-85530b4fbdc2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.275397 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdwrf\" (UniqueName: \"kubernetes.io/projected/4607c813-5202-460e-b583-a03dc13eee01-kube-api-access-tdwrf\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276108 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c279ea73-49f4-441d-9a36-bedd8388437e-node-pullsecrets\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276120 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fcz2\" (UniqueName: \"kubernetes.io/projected/c279ea73-49f4-441d-9a36-bedd8388437e-kube-api-access-4fcz2\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276224 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276269 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df306321-4c89-4909-834c-48ae131ac196-config\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276360 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-etcd-client\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276433 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-stats-auth\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276788 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c279ea73-49f4-441d-9a36-bedd8388437e-audit-dir\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276818 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfgvp\" (UniqueName: \"kubernetes.io/projected/7880dcf7-b6e3-4620-ba8b-013f207b02e0-kube-api-access-bfgvp\") pod \"ingress-canary-jsgfm\" (UID: \"7880dcf7-b6e3-4620-ba8b-013f207b02e0\") " pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276837 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w729g\" (UniqueName: \"kubernetes.io/projected/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-kube-api-access-w729g\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276924 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-default-certificate\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276962 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-client\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.276989 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-profile-collector-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.277014 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rktv7\" (UniqueName: \"kubernetes.io/projected/64c0a3f4-29f0-42fa-83bc-83e15171ed77-kube-api-access-rktv7\") pod \"auto-csr-approver-29550802-w82g5\" (UID: \"64c0a3f4-29f0-42fa-83bc-83e15171ed77\") " pod="openshift-infra/auto-csr-approver-29550802-w82g5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.277301 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.277388 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c279ea73-49f4-441d-9a36-bedd8388437e-audit-dir\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.277600 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-config\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278154 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-service-ca-bundle\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278222 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4063de5-2cdb-4d86-a74f-c98af0aae38d-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278403 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-image-import-ca\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278497 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278594 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cbcb6e-701d-45c4-8293-945716329cef-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278709 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d4063de5-2cdb-4d86-a74f-c98af0aae38d-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278741 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-registration-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278798 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dq54\" (UniqueName: \"kubernetes.io/projected/93ccefab-c47d-4725-a9ff-55cc50f678f0-kube-api-access-2dq54\") pod \"multus-admission-controller-857f4d67dd-vdjpl\" (UID: \"93ccefab-c47d-4725-a9ff-55cc50f678f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.278820 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:01.778807954 +0000 UTC m=+122.603543535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278843 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7bph\" (UniqueName: \"kubernetes.io/projected/385a2958-770a-4ce1-9b88-b897cb08c3eb-kube-api-access-q7bph\") pod \"package-server-manager-789f6589d5-4gfv7\" (UID: \"385a2958-770a-4ce1-9b88-b897cb08c3eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278897 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12f530da-08e1-42ee-a635-ca648ee4f5b3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278922 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8kkb\" (UniqueName: \"kubernetes.io/projected/9c16e62e-9744-45c8-88d9-0c289df743a1-kube-api-access-j8kkb\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278939 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfk64\" (UniqueName: \"kubernetes.io/projected/f96b54e5-804b-4d9e-9df9-538352c52073-kube-api-access-cfk64\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.278971 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcm6z\" (UniqueName: \"kubernetes.io/projected/974389aa-40cb-4191-a85b-d14e1aa175af-kube-api-access-vcm6z\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279029 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vtnd\" (UniqueName: \"kubernetes.io/projected/146e15b3-9b76-4a85-acb5-85530b4fbdc2-kube-api-access-6vtnd\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279064 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-bound-sa-token\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279088 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-config-volume\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279137 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279155 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-certificates\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279188 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946e899e-f89e-4f66-80a2-1f3effc06b7f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279205 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-audit\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279255 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96b54e5-804b-4d9e-9df9-538352c52073-config-volume\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279274 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/385a2958-770a-4ce1-9b88-b897cb08c3eb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4gfv7\" (UID: \"385a2958-770a-4ce1-9b88-b897cb08c3eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279292 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kqc7\" (UniqueName: \"kubernetes.io/projected/deb95cf2-78dc-43a2-acce-73f8be8e9f03-kube-api-access-8kqc7\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279312 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946e899e-f89e-4f66-80a2-1f3effc06b7f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279329 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb95cf2-78dc-43a2-acce-73f8be8e9f03-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279348 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ksmq\" (UniqueName: \"kubernetes.io/projected/07c9ac14-6e46-4243-a490-6467522ae8e4-kube-api-access-6ksmq\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279575 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-node-bootstrap-token\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279630 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddda8795-d770-4f73-a1c2-c988092f7fa4-config\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279656 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946e899e-f89e-4f66-80a2-1f3effc06b7f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279708 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-metrics-certs\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279730 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92h96\" (UniqueName: \"kubernetes.io/projected/700952b9-c6cb-403f-855b-b3cbd855e845-kube-api-access-92h96\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279762 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c9ac14-6e46-4243-a490-6467522ae8e4-serving-cert\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279829 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cbcb6e-701d-45c4-8293-945716329cef-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279850 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27kmp\" (UniqueName: \"kubernetes.io/projected/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-kube-api-access-27kmp\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279892 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-ca\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279922 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpkkd\" (UniqueName: \"kubernetes.io/projected/981360f1-a9de-4035-bbb0-2fae838061f5-kube-api-access-lpkkd\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279949 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-images\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.279965 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkb9w\" (UniqueName: \"kubernetes.io/projected/d4063de5-2cdb-4d86-a74f-c98af0aae38d-kube-api-access-rkb9w\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280059 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-trusted-ca\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280083 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-etcd-serving-ca\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280105 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-serving-cert\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280128 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/543c31b5-3ba9-4e04-aeaf-3ba7315da541-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280148 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t88ck\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-kube-api-access-t88ck\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280198 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv2qt\" (UniqueName: \"kubernetes.io/projected/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-kube-api-access-sv2qt\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280220 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280237 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d24t6\" (UniqueName: \"kubernetes.io/projected/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-kube-api-access-d24t6\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280256 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12f530da-08e1-42ee-a635-ca648ee4f5b3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280272 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/700952b9-c6cb-403f-855b-b3cbd855e845-webhook-cert\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.280288 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-certs\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.281386 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-certificates\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.284385 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-trusted-ca\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.291153 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.312478 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.332021 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.352664 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.371775 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.377217 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381396 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.381546 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:01.881522176 +0000 UTC m=+122.706257757 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381651 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df306321-4c89-4909-834c-48ae131ac196-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381687 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-srv-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381717 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-csi-data-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381747 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f530da-08e1-42ee-a635-ca648ee4f5b3-config\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381797 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c9ac14-6e46-4243-a490-6467522ae8e4-config\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381824 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/700952b9-c6cb-403f-855b-b3cbd855e845-apiservice-cert\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381855 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvrr9\" (UniqueName: \"kubernetes.io/projected/f93524c6-53a2-4df0-873b-241f5ff69a9d-kube-api-access-nvrr9\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381880 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/974389aa-40cb-4191-a85b-d14e1aa175af-proxy-tls\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381917 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bng8j\" (UID: \"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.381947 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsqmf\" (UniqueName: \"kubernetes.io/projected/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-kube-api-access-wsqmf\") pod \"control-plane-machine-set-operator-78cbb6b69f-bng8j\" (UID: \"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382011 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df306321-4c89-4909-834c-48ae131ac196-config\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382053 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-stats-auth\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382080 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfgvp\" (UniqueName: \"kubernetes.io/projected/7880dcf7-b6e3-4620-ba8b-013f207b02e0-kube-api-access-bfgvp\") pod \"ingress-canary-jsgfm\" (UID: \"7880dcf7-b6e3-4620-ba8b-013f207b02e0\") " pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382106 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w729g\" (UniqueName: \"kubernetes.io/projected/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-kube-api-access-w729g\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382134 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-default-certificate\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382188 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-profile-collector-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382219 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rktv7\" (UniqueName: \"kubernetes.io/projected/64c0a3f4-29f0-42fa-83bc-83e15171ed77-kube-api-access-rktv7\") pod \"auto-csr-approver-29550802-w82g5\" (UID: \"64c0a3f4-29f0-42fa-83bc-83e15171ed77\") " pod="openshift-infra/auto-csr-approver-29550802-w82g5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382266 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4063de5-2cdb-4d86-a74f-c98af0aae38d-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382309 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382344 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d4063de5-2cdb-4d86-a74f-c98af0aae38d-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382376 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-registration-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382412 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dq54\" (UniqueName: \"kubernetes.io/projected/93ccefab-c47d-4725-a9ff-55cc50f678f0-kube-api-access-2dq54\") pod \"multus-admission-controller-857f4d67dd-vdjpl\" (UID: \"93ccefab-c47d-4725-a9ff-55cc50f678f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382440 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7bph\" (UniqueName: \"kubernetes.io/projected/385a2958-770a-4ce1-9b88-b897cb08c3eb-kube-api-access-q7bph\") pod \"package-server-manager-789f6589d5-4gfv7\" (UID: \"385a2958-770a-4ce1-9b88-b897cb08c3eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382468 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12f530da-08e1-42ee-a635-ca648ee4f5b3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382496 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8kkb\" (UniqueName: \"kubernetes.io/projected/9c16e62e-9744-45c8-88d9-0c289df743a1-kube-api-access-j8kkb\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382524 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfk64\" (UniqueName: \"kubernetes.io/projected/f96b54e5-804b-4d9e-9df9-538352c52073-kube-api-access-cfk64\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382552 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcm6z\" (UniqueName: \"kubernetes.io/projected/974389aa-40cb-4191-a85b-d14e1aa175af-kube-api-access-vcm6z\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382693 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-config-volume\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382759 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96b54e5-804b-4d9e-9df9-538352c52073-config-volume\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382792 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/385a2958-770a-4ce1-9b88-b897cb08c3eb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4gfv7\" (UID: \"385a2958-770a-4ce1-9b88-b897cb08c3eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382821 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kqc7\" (UniqueName: \"kubernetes.io/projected/deb95cf2-78dc-43a2-acce-73f8be8e9f03-kube-api-access-8kqc7\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382855 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb95cf2-78dc-43a2-acce-73f8be8e9f03-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382885 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4063de5-2cdb-4d86-a74f-c98af0aae38d-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382884 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ksmq\" (UniqueName: \"kubernetes.io/projected/07c9ac14-6e46-4243-a490-6467522ae8e4-kube-api-access-6ksmq\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.382979 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-node-bootstrap-token\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383012 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-metrics-certs\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383032 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92h96\" (UniqueName: \"kubernetes.io/projected/700952b9-c6cb-403f-855b-b3cbd855e845-kube-api-access-92h96\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383054 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c9ac14-6e46-4243-a490-6467522ae8e4-serving-cert\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383079 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27kmp\" (UniqueName: \"kubernetes.io/projected/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-kube-api-access-27kmp\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383110 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpkkd\" (UniqueName: \"kubernetes.io/projected/981360f1-a9de-4035-bbb0-2fae838061f5-kube-api-access-lpkkd\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383132 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-images\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.383254 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:01.883245839 +0000 UTC m=+122.707981420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383292 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/543c31b5-3ba9-4e04-aeaf-3ba7315da541-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383321 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkb9w\" (UniqueName: \"kubernetes.io/projected/d4063de5-2cdb-4d86-a74f-c98af0aae38d-kube-api-access-rkb9w\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383366 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12f530da-08e1-42ee-a635-ca648ee4f5b3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383384 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/700952b9-c6cb-403f-855b-b3cbd855e845-webhook-cert\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383401 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-certs\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383472 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-plugins-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383495 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/d4063de5-2cdb-4d86-a74f-c98af0aae38d-ready\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383516 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7880dcf7-b6e3-4620-ba8b-013f207b02e0-cert\") pod \"ingress-canary-jsgfm\" (UID: \"7880dcf7-b6e3-4620-ba8b-013f207b02e0\") " pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383534 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf7zs\" (UniqueName: \"kubernetes.io/projected/03b56c7f-0613-419a-8a32-fe678a4a3e93-kube-api-access-gf7zs\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383561 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/543c31b5-3ba9-4e04-aeaf-3ba7315da541-proxy-tls\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383590 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/700952b9-c6cb-403f-855b-b3cbd855e845-tmpfs\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383607 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f93524c6-53a2-4df0-873b-241f5ff69a9d-signing-key\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383624 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383656 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383673 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-socket-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383696 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93ccefab-c47d-4725-a9ff-55cc50f678f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vdjpl\" (UID: \"93ccefab-c47d-4725-a9ff-55cc50f678f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383723 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383740 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-srv-cert\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383762 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df306321-4c89-4909-834c-48ae131ac196-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383778 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbw5p\" (UniqueName: \"kubernetes.io/projected/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-kube-api-access-sbw5p\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383812 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c86z\" (UniqueName: \"kubernetes.io/projected/543c31b5-3ba9-4e04-aeaf-3ba7315da541-kube-api-access-8c86z\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383828 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-metrics-tls\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383845 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f93524c6-53a2-4df0-873b-241f5ff69a9d-signing-cabundle\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383862 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb95cf2-78dc-43a2-acce-73f8be8e9f03-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383880 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzd4h\" (UniqueName: \"kubernetes.io/projected/cb17c405-b001-4e0a-8721-776ce83db499-kube-api-access-qzd4h\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383896 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383934 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383951 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb17c405-b001-4e0a-8721-776ce83db499-service-ca-bundle\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383966 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-mountpoint-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.383964 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-config-volume\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.384803 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07c9ac14-6e46-4243-a490-6467522ae8e4-config\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.385036 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-csi-data-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.385731 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-registration-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.386040 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d4063de5-2cdb-4d86-a74f-c98af0aae38d-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.386316 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-mountpoint-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.387241 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb17c405-b001-4e0a-8721-776ce83db499-service-ca-bundle\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.387331 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/d4063de5-2cdb-4d86-a74f-c98af0aae38d-ready\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.387426 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f93524c6-53a2-4df0-873b-241f5ff69a9d-signing-cabundle\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.387698 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/700952b9-c6cb-403f-855b-b3cbd855e845-tmpfs\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.387727 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-default-certificate\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.388271 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96b54e5-804b-4d9e-9df9-538352c52073-config-volume\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.388370 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/deb95cf2-78dc-43a2-acce-73f8be8e9f03-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.387426 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-plugins-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.388580 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/700952b9-c6cb-403f-855b-b3cbd855e845-apiservice-cert\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.388873 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/03b56c7f-0613-419a-8a32-fe678a4a3e93-socket-dir\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.388999 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.389145 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/543c31b5-3ba9-4e04-aeaf-3ba7315da541-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.389831 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-node-bootstrap-token\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.391481 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/700952b9-c6cb-403f-855b-b3cbd855e845-webhook-cert\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.391641 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-srv-cert\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.392002 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f93524c6-53a2-4df0-873b-241f5ff69a9d-signing-key\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.392111 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-stats-auth\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.394228 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-certs\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.394519 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7880dcf7-b6e3-4620-ba8b-013f207b02e0-cert\") pod \"ingress-canary-jsgfm\" (UID: \"7880dcf7-b6e3-4620-ba8b-013f207b02e0\") " pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.394553 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/385a2958-770a-4ce1-9b88-b897cb08c3eb-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-4gfv7\" (UID: \"385a2958-770a-4ce1-9b88-b897cb08c3eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.396883 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07c9ac14-6e46-4243-a490-6467522ae8e4-serving-cert\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.397734 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-metrics-tls\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.398019 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.402134 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/deb95cf2-78dc-43a2-acce-73f8be8e9f03-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.402402 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb17c405-b001-4e0a-8721-776ce83db499-metrics-certs\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.432012 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.442341 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfwlq\" (UniqueName: \"kubernetes.io/projected/5537a456-f771-4a55-b811-f20b50cd6446-kube-api-access-kfwlq\") pod \"oauth-openshift-558db77b4-pbl6c\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.464413 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.465257 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-service-ca\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.480599 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.483908 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-config\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.484585 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.484825 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:01.984799204 +0000 UTC m=+122.809534785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.485221 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.485863 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:01.985856457 +0000 UTC m=+122.810592038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.516344 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.522766 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-serving-cert\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.541010 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.545087 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ddda8795-d770-4f73-a1c2-c988092f7fa4-auth-proxy-config\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.556099 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.566648 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-serving-cert\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.572968 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.577921 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ddda8795-d770-4f73-a1c2-c988092f7fa4-machine-approver-tls\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.584086 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb"] Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.586857 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.587028 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.08700822 +0000 UTC m=+122.911743801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.587484 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.587923 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.087915758 +0000 UTC m=+122.912651349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.633332 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.637792 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/146e15b3-9b76-4a85-acb5-85530b4fbdc2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.651447 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.655433 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-config\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.671978 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.678502 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-serving-cert\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.689780 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.690395 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.190363322 +0000 UTC m=+123.015099063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.690916 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.692015 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.192003692 +0000 UTC m=+123.016739273 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.699007 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.706200 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-trusted-ca\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.736377 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.747240 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/146e15b3-9b76-4a85-acb5-85530b4fbdc2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.752439 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.757163 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-config\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.762864 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" event={"ID":"ca45f39c-f9bd-4115-83f4-e866533d1090","Type":"ContainerStarted","Data":"45da4bd8cbd8825b256893a0fa7504b8312fd4d9ec75df336a62a69a5ecb7bb1"} Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.764087 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" event={"ID":"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758","Type":"ContainerStarted","Data":"2626b7d3fabe7e9968f112a4035550f88f89b451b0ebeacc51e6417569d79370"} Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.764121 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" event={"ID":"0a58b7e4-7d03-4cc1-8c87-f1a0ed51f758","Type":"ContainerStarted","Data":"e3a573f3b535fbcf5a0299a0700d465c5b7cca62497ea5f04baf1f5ca34807c5"} Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.768700 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.772749 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.781561 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-encryption-config\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.792353 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.792579 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.292549297 +0000 UTC m=+123.117284878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.792965 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.793381 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.293372742 +0000 UTC m=+123.118108533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.812220 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.820391 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97ncr\" (UniqueName: \"kubernetes.io/projected/553737d8-1213-41ef-a0ee-d1f85d749634-kube-api-access-97ncr\") pod \"cluster-samples-operator-665b6dd947-tr7zk\" (UID: \"553737d8-1213-41ef-a0ee-d1f85d749634\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.833927 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.842576 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4607c813-5202-460e-b583-a03dc13eee01-serving-cert\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.877794 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.892791 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-tls\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.896734 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.896906 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.396881888 +0000 UTC m=+123.221617469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.897631 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.897988 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.397978621 +0000 UTC m=+123.222714202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.906878 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.912680 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c279ea73-49f4-441d-9a36-bedd8388437e-etcd-client\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.912781 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.921776 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-client\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.931831 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.939180 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-config\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.952027 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.959791 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-service-ca-bundle\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.972455 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.979877 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-image-import-ca\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.992652 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.999055 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.999268 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.499233498 +0000 UTC m=+123.323969079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:01 crc kubenswrapper[4692]: I0309 09:22:01.999391 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:01 crc kubenswrapper[4692]: E0309 09:22:01.999741 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.499730303 +0000 UTC m=+123.324465884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.005631 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64cbcb6e-701d-45c4-8293-945716329cef-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.037282 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-bound-sa-token\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.047948 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.051231 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.060916 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-audit\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.072204 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.084510 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946e899e-f89e-4f66-80a2-1f3effc06b7f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.093212 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.100814 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.100996 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.60096828 +0000 UTC m=+123.425703861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.101248 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.101854 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.601840236 +0000 UTC m=+123.426575817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.102844 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64cbcb6e-701d-45c4-8293-945716329cef-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.114341 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.124505 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-etcd-ca\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.149729 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t88ck\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-kube-api-access-t88ck\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.176861 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.184741 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.192146 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.197686 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c279ea73-49f4-441d-9a36-bedd8388437e-etcd-serving-ca\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.202778 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.203007 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.702976069 +0000 UTC m=+123.527711670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.203238 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.203539 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.703529386 +0000 UTC m=+123.528265167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.216160 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.228520 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-serving-cert\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.251806 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.261933 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddda8795-d770-4f73-a1c2-c988092f7fa4-config\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.281358 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.283073 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4607c813-5202-460e-b583-a03dc13eee01-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.291782 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.299420 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc559\" (UniqueName: \"kubernetes.io/projected/7b8b20b2-f067-4abd-adc3-7b1a2bf36742-kube-api-access-qc559\") pod \"console-f9d7485db-rddqb\" (UID: \"7b8b20b2-f067-4abd-adc3-7b1a2bf36742\") " pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.301132 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mwqz\" (UniqueName: \"kubernetes.io/projected/aa908134-de3c-43e0-ace1-94fc548861c8-kube-api-access-5mwqz\") pod \"downloads-7954f5f757-snq8b\" (UID: \"aa908134-de3c-43e0-ace1-94fc548861c8\") " pod="openshift-console/downloads-7954f5f757-snq8b" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.304405 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.304507 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.804484704 +0000 UTC m=+123.629220285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.304978 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.305332 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.80532226 +0000 UTC m=+123.630057841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.311521 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.320147 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph2ts\" (UniqueName: \"kubernetes.io/projected/d59db20e-6720-4cc9-bfd8-b3477c7298cf-kube-api-access-ph2ts\") pod \"dns-operator-744455d44c-rfkdj\" (UID: \"d59db20e-6720-4cc9-bfd8-b3477c7298cf\") " pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.332345 4692 projected.go:288] Couldn't get configMap openshift-ingress-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.332389 4692 projected.go:194] Error preparing data for projected volume kube-api-access-lrpmz for pod openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.332454 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-kube-api-access-lrpmz podName:8385d993-0021-4066-a9c6-57bf7ee4be33 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.832433099 +0000 UTC m=+123.657168680 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-lrpmz" (UniqueName: "kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-kube-api-access-lrpmz") pod "ingress-operator-5b745b69d9-9w27p" (UID: "8385d993-0021-4066-a9c6-57bf7ee4be33") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.332452 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.352697 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.359521 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmvx4\" (UniqueName: \"kubernetes.io/projected/d7e9d340-f169-49ff-82ae-21089bdb28f6-kube-api-access-lmvx4\") pod \"openshift-apiserver-operator-796bbdcf4f-l8794\" (UID: \"d7e9d340-f169-49ff-82ae-21089bdb28f6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.383834 4692 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.383925 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-images podName:974389aa-40cb-4191-a85b-d14e1aa175af nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.883904013 +0000 UTC m=+123.708639594 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-images") pod "machine-config-operator-74547568cd-ktw79" (UID: "974389aa-40cb-4191-a85b-d14e1aa175af") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.383929 4692 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.383986 4692 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.383992 4692 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.384019 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/974389aa-40cb-4191-a85b-d14e1aa175af-proxy-tls podName:974389aa-40cb-4191-a85b-d14e1aa175af nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.883998826 +0000 UTC m=+123.708734407 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/974389aa-40cb-4191-a85b-d14e1aa175af-proxy-tls") pod "machine-config-operator-74547568cd-ktw79" (UID: "974389aa-40cb-4191-a85b-d14e1aa175af") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.383933 4692 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.384135 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/12f530da-08e1-42ee-a635-ca648ee4f5b3-config podName:12f530da-08e1-42ee-a635-ca648ee4f5b3 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.884108729 +0000 UTC m=+123.708844510 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/12f530da-08e1-42ee-a635-ca648ee4f5b3-config") pod "kube-controller-manager-operator-78b949d7b-xsdcf" (UID: "12f530da-08e1-42ee-a635-ca648ee4f5b3") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.384155 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-control-plane-machine-set-operator-tls podName:14a9e924-9e6f-4a5e-82f4-ea2b52ae4334 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.88414549 +0000 UTC m=+123.708881301 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-bng8j" (UID: "14a9e924-9e6f-4a5e-82f4-ea2b52ae4334") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.384206 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/df306321-4c89-4909-834c-48ae131ac196-config podName:df306321-4c89-4909-834c-48ae131ac196 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.884193912 +0000 UTC m=+123.708929693 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/df306321-4c89-4909-834c-48ae131ac196-config") pod "kube-apiserver-operator-766d6c64bb-7n6dg" (UID: "df306321-4c89-4909-834c-48ae131ac196") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.385218 4692 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.385269 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics podName:96013f1f-d2b0-42ec-a4f3-3626f48be8e1 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.885250954 +0000 UTC m=+123.709986535 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics") pod "marketplace-operator-79b997595-rdvhn" (UID: "96013f1f-d2b0-42ec-a4f3-3626f48be8e1") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.385219 4692 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.385300 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-profile-collector-cert podName:981360f1-a9de-4035-bbb0-2fae838061f5 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.885295075 +0000 UTC m=+123.710030656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-profile-collector-cert") pod "catalog-operator-68c6474976-26mfz" (UID: "981360f1-a9de-4035-bbb0-2fae838061f5") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.386227 4692 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.386257 4692 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.386275 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12f530da-08e1-42ee-a635-ca648ee4f5b3-serving-cert podName:12f530da-08e1-42ee-a635-ca648ee4f5b3 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.886264305 +0000 UTC m=+123.710999886 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/12f530da-08e1-42ee-a635-ca648ee4f5b3-serving-cert") pod "kube-controller-manager-operator-78b949d7b-xsdcf" (UID: "12f530da-08e1-42ee-a635-ca648ee4f5b3") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.386294 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-srv-cert podName:981360f1-a9de-4035-bbb0-2fae838061f5 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.886285396 +0000 UTC m=+123.711020977 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-srv-cert") pod "catalog-operator-68c6474976-26mfz" (UID: "981360f1-a9de-4035-bbb0-2fae838061f5") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.387381 4692 secret.go:188] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.387390 4692 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.387442 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/df306321-4c89-4909-834c-48ae131ac196-serving-cert podName:df306321-4c89-4909-834c-48ae131ac196 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.887427611 +0000 UTC m=+123.712163362 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/df306321-4c89-4909-834c-48ae131ac196-serving-cert") pod "kube-apiserver-operator-766d6c64bb-7n6dg" (UID: "df306321-4c89-4909-834c-48ae131ac196") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.387459 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/543c31b5-3ba9-4e04-aeaf-3ba7315da541-proxy-tls podName:543c31b5-3ba9-4e04-aeaf-3ba7315da541 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.887451531 +0000 UTC m=+123.712187312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/543c31b5-3ba9-4e04-aeaf-3ba7315da541-proxy-tls") pod "machine-config-controller-84d6567774-87j6g" (UID: "543c31b5-3ba9-4e04-aeaf-3ba7315da541") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.388667 4692 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.388718 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca podName:96013f1f-d2b0-42ec-a4f3-3626f48be8e1 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.88870503 +0000 UTC m=+123.713440801 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca") pod "marketplace-operator-79b997595-rdvhn" (UID: "96013f1f-d2b0-42ec-a4f3-3626f48be8e1") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.389749 4692 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.389770 4692 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.389799 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume podName:f96b54e5-804b-4d9e-9df9-538352c52073 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.889787763 +0000 UTC m=+123.714523534 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-volume" (UniqueName: "kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume") pod "collect-profiles-29550795-5rfpv" (UID: "f96b54e5-804b-4d9e-9df9-538352c52073") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.389815 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/93ccefab-c47d-4725-a9ff-55cc50f678f0-webhook-certs podName:93ccefab-c47d-4725-a9ff-55cc50f678f0 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.889804803 +0000 UTC m=+123.714540584 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/93ccefab-c47d-4725-a9ff-55cc50f678f0-webhook-certs") pod "multus-admission-controller-857f4d67dd-vdjpl" (UID: "93ccefab-c47d-4725-a9ff-55cc50f678f0") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.389826 4692 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/pprof-cert: failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.389853 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-profile-collector-cert podName:9c16e62e-9744-45c8-88d9-0c289df743a1 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.889844775 +0000 UTC m=+123.714580566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "profile-collector-cert" (UniqueName: "kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-profile-collector-cert") pod "olm-operator-6b444d44fb-29jhq" (UID: "9c16e62e-9744-45c8-88d9-0c289df743a1") : failed to sync secret cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.390937 4692 request.go:700] Waited for 1.007871915s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/serviceaccounts/olm-operator-serviceaccount/token Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.404707 4692 projected.go:288] Couldn't get configMap openshift-kube-storage-version-migrator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.404749 4692 projected.go:194] Error preparing data for projected volume kube-api-access-wj8pl for pod openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.404830 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f73e489b-1a01-4608-a493-e0ce7dd2c2c5-kube-api-access-wj8pl podName:f73e489b-1a01-4608-a493-e0ce7dd2c2c5 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.904807492 +0000 UTC m=+123.729543073 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-wj8pl" (UniqueName: "kubernetes.io/projected/f73e489b-1a01-4608-a493-e0ce7dd2c2c5-kube-api-access-wj8pl") pod "migrator-59844c95c7-ftqvs" (UID: "f73e489b-1a01-4608-a493-e0ce7dd2c2c5") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.406562 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.407261 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:02.907244307 +0000 UTC m=+123.731979888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.411082 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.465456 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcm6z\" (UniqueName: \"kubernetes.io/projected/974389aa-40cb-4191-a85b-d14e1aa175af-kube-api-access-vcm6z\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.487115 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfgvp\" (UniqueName: \"kubernetes.io/projected/7880dcf7-b6e3-4620-ba8b-013f207b02e0-kube-api-access-bfgvp\") pod \"ingress-canary-jsgfm\" (UID: \"7880dcf7-b6e3-4620-ba8b-013f207b02e0\") " pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.507068 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvrr9\" (UniqueName: \"kubernetes.io/projected/f93524c6-53a2-4df0-873b-241f5ff69a9d-kube-api-access-nvrr9\") pod \"service-ca-9c57cc56f-rxzs8\" (UID: \"f93524c6-53a2-4df0-873b-241f5ff69a9d\") " pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.508121 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.508476 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.008459762 +0000 UTC m=+123.833195523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.514813 4692 projected.go:288] Couldn't get configMap openshift-cluster-machine-approver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.532153 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w729g\" (UniqueName: \"kubernetes.io/projected/51b39b63-d3ad-4a29-b7b5-6cede45e2fa5-kube-api-access-w729g\") pod \"dns-default-vbp6d\" (UID: \"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5\") " pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.532418 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.551505 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsqmf\" (UniqueName: \"kubernetes.io/projected/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-kube-api-access-wsqmf\") pod \"control-plane-machine-set-operator-78cbb6b69f-bng8j\" (UID: \"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.565881 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kqc7\" (UniqueName: \"kubernetes.io/projected/deb95cf2-78dc-43a2-acce-73f8be8e9f03-kube-api-access-8kqc7\") pod \"kube-storage-version-migrator-operator-b67b599dd-ldfbc\" (UID: \"deb95cf2-78dc-43a2-acce-73f8be8e9f03\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.571945 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.578416 4692 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.578505 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.592211 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.608246 4692 projected.go:288] Couldn't get configMap openshift-etcd-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.609912 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.610772 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.110754431 +0000 UTC m=+123.935490012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.625134 4692 projected.go:288] Couldn't get configMap openshift-kube-scheduler-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.625182 4692 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.625238 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/64cbcb6e-701d-45c4-8293-945716329cef-kube-api-access podName:64cbcb6e-701d-45c4-8293-945716329cef nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.125219863 +0000 UTC m=+123.949955444 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/64cbcb6e-701d-45c4-8293-945716329cef-kube-api-access") pod "openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" (UID: "64cbcb6e-701d-45c4-8293-945716329cef") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.632372 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ksmq\" (UniqueName: \"kubernetes.io/projected/07c9ac14-6e46-4243-a490-6467522ae8e4-kube-api-access-6ksmq\") pod \"service-ca-operator-777779d784-4qpm7\" (UID: \"07c9ac14-6e46-4243-a490-6467522ae8e4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.669921 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rktv7\" (UniqueName: \"kubernetes.io/projected/64c0a3f4-29f0-42fa-83bc-83e15171ed77-kube-api-access-rktv7\") pod \"auto-csr-approver-29550802-w82g5\" (UID: \"64c0a3f4-29f0-42fa-83bc-83e15171ed77\") " pod="openshift-infra/auto-csr-approver-29550802-w82g5" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.677462 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.692401 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.714524 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.714985 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.214970008 +0000 UTC m=+124.039705589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.720434 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.722524 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.727059 4692 projected.go:288] Couldn't get configMap openshift-authentication-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.731975 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.734590 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.748486 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jsgfm" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.748673 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vbp6d"] Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.776921 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkb9w\" (UniqueName: \"kubernetes.io/projected/d4063de5-2cdb-4d86-a74f-c98af0aae38d-kube-api-access-rkb9w\") pod \"cni-sysctl-allowlist-ds-c9rx4\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:02 crc kubenswrapper[4692]: W0309 09:22:02.778593 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51b39b63_d3ad_4a29_b7b5_6cede45e2fa5.slice/crio-cd0fec2bc09f98f7c72a467823d13297a568ec82d0b1260cee6716fa3370fb45 WatchSource:0}: Error finding container cd0fec2bc09f98f7c72a467823d13297a568ec82d0b1260cee6716fa3370fb45: Status 404 returned error can't find the container with id cd0fec2bc09f98f7c72a467823d13297a568ec82d0b1260cee6716fa3370fb45 Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.784367 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.795757 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dq54\" (UniqueName: \"kubernetes.io/projected/93ccefab-c47d-4725-a9ff-55cc50f678f0-kube-api-access-2dq54\") pod \"multus-admission-controller-857f4d67dd-vdjpl\" (UID: \"93ccefab-c47d-4725-a9ff-55cc50f678f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.814002 4692 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.815378 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.815537 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.315520033 +0000 UTC m=+124.140255614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.815888 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.816964 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.316947487 +0000 UTC m=+124.141683298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.849614 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.856152 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf7zs\" (UniqueName: \"kubernetes.io/projected/03b56c7f-0613-419a-8a32-fe678a4a3e93-kube-api-access-gf7zs\") pod \"csi-hostpathplugin-jdvnz\" (UID: \"03b56c7f-0613-419a-8a32-fe678a4a3e93\") " pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.871107 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pbl6c"] Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.876896 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzd4h\" (UniqueName: \"kubernetes.io/projected/cb17c405-b001-4e0a-8721-776ce83db499-kube-api-access-qzd4h\") pod \"router-default-5444994796-d5xx5\" (UID: \"cb17c405-b001-4e0a-8721-776ce83db499\") " pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.879619 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.898134 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550802-w82g5" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.911859 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.919158 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.919410 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-images\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.919470 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12f530da-08e1-42ee-a635-ca648ee4f5b3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.919496 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.419466253 +0000 UTC m=+124.244201824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.920155 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/974389aa-40cb-4191-a85b-d14e1aa175af-images\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931340 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/543c31b5-3ba9-4e04-aeaf-3ba7315da541-proxy-tls\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931397 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931419 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931449 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93ccefab-c47d-4725-a9ff-55cc50f678f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vdjpl\" (UID: \"93ccefab-c47d-4725-a9ff-55cc50f678f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931474 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931494 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj8pl\" (UniqueName: \"kubernetes.io/projected/f73e489b-1a01-4608-a493-e0ce7dd2c2c5-kube-api-access-wj8pl\") pod \"migrator-59844c95c7-ftqvs\" (UID: \"f73e489b-1a01-4608-a493-e0ce7dd2c2c5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931511 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df306321-4c89-4909-834c-48ae131ac196-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931575 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931600 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-srv-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931619 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f530da-08e1-42ee-a635-ca648ee4f5b3-config\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931644 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrpmz\" (UniqueName: \"kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-kube-api-access-lrpmz\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931661 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/974389aa-40cb-4191-a85b-d14e1aa175af-proxy-tls\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931680 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bng8j\" (UID: \"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931714 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df306321-4c89-4909-834c-48ae131ac196-config\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931740 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-profile-collector-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.931770 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:02 crc kubenswrapper[4692]: E0309 09:22:02.932096 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.432085088 +0000 UTC m=+124.256820669 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.937085 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12f530da-08e1-42ee-a635-ca648ee4f5b3-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.937199 4692 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.937237 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.941754 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12f530da-08e1-42ee-a635-ca648ee4f5b3-config\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.942866 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.943356 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.944126 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/974389aa-40cb-4191-a85b-d14e1aa175af-proxy-tls\") pod \"machine-config-operator-74547568cd-ktw79\" (UID: \"974389aa-40cb-4191-a85b-d14e1aa175af\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.945112 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df306321-4c89-4909-834c-48ae131ac196-config\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.945110 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.945297 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/14a9e924-9e6f-4a5e-82f4-ea2b52ae4334-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bng8j\" (UID: \"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.948921 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrpmz\" (UniqueName: \"kubernetes.io/projected/8385d993-0021-4066-a9c6-57bf7ee4be33-kube-api-access-lrpmz\") pod \"ingress-operator-5b745b69d9-9w27p\" (UID: \"8385d993-0021-4066-a9c6-57bf7ee4be33\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.959599 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-profile-collector-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.972903 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9c16e62e-9744-45c8-88d9-0c289df743a1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.977121 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.988771 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c86z\" (UniqueName: \"kubernetes.io/projected/543c31b5-3ba9-4e04-aeaf-3ba7315da541-kube-api-access-8c86z\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:02 crc kubenswrapper[4692]: I0309 09:22:02.991849 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/93ccefab-c47d-4725-a9ff-55cc50f678f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-vdjpl\" (UID: \"93ccefab-c47d-4725-a9ff-55cc50f678f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.013770 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.015872 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.020441 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/543c31b5-3ba9-4e04-aeaf-3ba7315da541-proxy-tls\") pod \"machine-config-controller-84d6567774-87j6g\" (UID: \"543c31b5-3ba9-4e04-aeaf-3ba7315da541\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.034254 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.034391 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.534359516 +0000 UTC m=+124.359095097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.034481 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.034968 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.534956155 +0000 UTC m=+124.359691736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.041464 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df306321-4c89-4909-834c-48ae131ac196-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.049913 4692 projected.go:288] Couldn't get configMap openshift-controller-manager-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.057565 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.062791 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27kmp\" (UniqueName: \"kubernetes.io/projected/fbcc5285-0190-40e3-968b-2ff8de9bb6a9-kube-api-access-27kmp\") pod \"machine-config-server-fct6h\" (UID: \"fbcc5285-0190-40e3-968b-2ff8de9bb6a9\") " pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.078020 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/981360f1-a9de-4035-bbb0-2fae838061f5-srv-cert\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.085201 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.086500 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc"] Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.092216 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj8pl\" (UniqueName: \"kubernetes.io/projected/f73e489b-1a01-4608-a493-e0ce7dd2c2c5-kube-api-access-wj8pl\") pod \"migrator-59844c95c7-ftqvs\" (UID: \"f73e489b-1a01-4608-a493-e0ce7dd2c2c5\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.115398 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.120144 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.122998 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.134726 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.135878 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.136212 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/64cbcb6e-701d-45c4-8293-945716329cef-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.136554 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.636540412 +0000 UTC m=+124.461275993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.146424 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fct6h" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.152917 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.163638 4692 projected.go:288] Couldn't get configMap openshift-config-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: W0309 09:22:03.170322 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddeb95cf2_78dc_43a2_acce_73f8be8e9f03.slice/crio-173a82f3ba571b0d7ac1fa29573710d0549ef200f584a43ab83992e57bda9c20 WatchSource:0}: Error finding container 173a82f3ba571b0d7ac1fa29573710d0549ef200f584a43ab83992e57bda9c20: Status 404 returned error can't find the container with id 173a82f3ba571b0d7ac1fa29573710d0549ef200f584a43ab83992e57bda9c20 Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.181403 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.193475 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jsgfm"] Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.196883 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.200592 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/64cbcb6e-701d-45c4-8293-945716329cef-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-mpwzq\" (UID: \"64cbcb6e-701d-45c4-8293-945716329cef\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.215631 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.234732 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.237750 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.238197 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.73818522 +0000 UTC m=+124.562920801 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.254072 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7"] Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.255754 4692 projected.go:288] Couldn't get configMap openshift-console-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.256243 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.276147 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: W0309 09:22:03.294019 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07c9ac14_6e46_4243_a490_6467522ae8e4.slice/crio-66b4dff37343ca63bef4851dccb3e3be0ad9006c41f981b43222180c64b05847 WatchSource:0}: Error finding container 66b4dff37343ca63bef4851dccb3e3be0ad9006c41f981b43222180c64b05847: Status 404 returned error can't find the container with id 66b4dff37343ca63bef4851dccb3e3be0ad9006c41f981b43222180c64b05847 Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.299661 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.329058 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.336643 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.339099 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.339563 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.339795 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.839756517 +0000 UTC m=+124.664492098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.339974 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.345153 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68ff6f00-0ac5-41d9-b762-933944bdb507-metrics-certs\") pod \"network-metrics-daemon-9dn4z\" (UID: \"68ff6f00-0ac5-41d9-b762-933944bdb507\") " pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.346452 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.846436731 +0000 UTC m=+124.671172312 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.347396 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12f530da-08e1-42ee-a635-ca648ee4f5b3-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-xsdcf\" (UID: \"12f530da-08e1-42ee-a635-ca648ee4f5b3\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.350981 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfk64\" (UniqueName: \"kubernetes.io/projected/f96b54e5-804b-4d9e-9df9-538352c52073-kube-api-access-cfk64\") pod \"collect-profiles-29550795-5rfpv\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.352102 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7bph\" (UniqueName: \"kubernetes.io/projected/385a2958-770a-4ce1-9b88-b897cb08c3eb-kube-api-access-q7bph\") pod \"package-server-manager-789f6589d5-4gfv7\" (UID: \"385a2958-770a-4ce1-9b88-b897cb08c3eb\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.352360 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8kkb\" (UniqueName: \"kubernetes.io/projected/9c16e62e-9744-45c8-88d9-0c289df743a1-kube-api-access-j8kkb\") pod \"olm-operator-6b444d44fb-29jhq\" (UID: \"9c16e62e-9744-45c8-88d9-0c289df743a1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.352485 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.354858 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.357628 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92h96\" (UniqueName: \"kubernetes.io/projected/700952b9-c6cb-403f-855b-b3cbd855e845-kube-api-access-92h96\") pod \"packageserver-d55dfcdfc-kn54j\" (UID: \"700952b9-c6cb-403f-855b-b3cbd855e845\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.359776 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk"] Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.366914 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpkkd\" (UniqueName: \"kubernetes.io/projected/981360f1-a9de-4035-bbb0-2fae838061f5-kube-api-access-lpkkd\") pod \"catalog-operator-68c6474976-26mfz\" (UID: \"981360f1-a9de-4035-bbb0-2fae838061f5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.371350 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.375880 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.382492 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.397976 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.405154 4692 projected.go:194] Error preparing data for projected volume kube-api-access-8nzq9 for pod openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.405263 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ddda8795-d770-4f73-a1c2-c988092f7fa4-kube-api-access-8nzq9 podName:ddda8795-d770-4f73-a1c2-c988092f7fa4 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.90524007 +0000 UTC m=+124.729975651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-8nzq9" (UniqueName: "kubernetes.io/projected/ddda8795-d770-4f73-a1c2-c988092f7fa4-kube-api-access-8nzq9") pod "machine-approver-56656f9798-nvv5s" (UID: "ddda8795-d770-4f73-a1c2-c988092f7fa4") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.412517 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.414507 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9dn4z" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.421339 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-snq8b" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.433231 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.439689 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.440370 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.440824 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.940811507 +0000 UTC m=+124.765547088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.451795 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.464066 4692 projected.go:194] Error preparing data for projected volume kube-api-access-bm5sx for pod openshift-etcd-operator/etcd-operator-b45778765-9vg42: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.464157 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-kube-api-access-bm5sx podName:1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.964135441 +0000 UTC m=+124.788871022 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-bm5sx" (UniqueName: "kubernetes.io/projected/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-kube-api-access-bm5sx") pod "etcd-operator-b45778765-9vg42" (UID: "1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.474691 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.482384 4692 projected.go:194] Error preparing data for projected volume kube-api-access-tdwrf for pod openshift-authentication-operator/authentication-operator-69f744f599-qpk8p: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.482455 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4607c813-5202-460e-b583-a03dc13eee01-kube-api-access-tdwrf podName:4607c813-5202-460e-b583-a03dc13eee01 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.982436501 +0000 UTC m=+124.807172082 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-tdwrf" (UniqueName: "kubernetes.io/projected/4607c813-5202-460e-b583-a03dc13eee01-kube-api-access-tdwrf") pod "authentication-operator-69f744f599-qpk8p" (UID: "4607c813-5202-460e-b583-a03dc13eee01") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.490751 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.495290 4692 projected.go:194] Error preparing data for projected volume kube-api-access-4fcz2 for pod openshift-apiserver/apiserver-76f77b778f-4sdrj: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.495362 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c279ea73-49f4-441d-9a36-bedd8388437e-kube-api-access-4fcz2 podName:c279ea73-49f4-441d-9a36-bedd8388437e nodeName:}" failed. No retries permitted until 2026-03-09 09:22:03.995344535 +0000 UTC m=+124.820080116 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-4fcz2" (UniqueName: "kubernetes.io/projected/c279ea73-49f4-441d-9a36-bedd8388437e-kube-api-access-4fcz2") pod "apiserver-76f77b778f-4sdrj" (UID: "c279ea73-49f4-441d-9a36-bedd8388437e") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.510348 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550802-w82g5"] Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.512508 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.531728 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rxzs8"] Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.535937 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.536665 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/df306321-4c89-4909-834c-48ae131ac196-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7n6dg\" (UID: \"df306321-4c89-4909-834c-48ae131ac196\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.536758 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.541755 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.542001 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.041988582 +0000 UTC m=+124.866724173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.551661 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.572616 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.576473 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.580123 4692 projected.go:194] Error preparing data for projected volume kube-api-access-6vtnd for pod openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.580201 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/146e15b3-9b76-4a85-acb5-85530b4fbdc2-kube-api-access-6vtnd podName:146e15b3-9b76-4a85-acb5-85530b4fbdc2 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.08018304 +0000 UTC m=+124.904918621 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-6vtnd" (UniqueName: "kubernetes.io/projected/146e15b3-9b76-4a85-acb5-85530b4fbdc2-kube-api-access-6vtnd") pod "openshift-controller-manager-operator-756b6f6bc6-qjf2f" (UID: "146e15b3-9b76-4a85-acb5-85530b4fbdc2") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.617373 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.620904 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-jdvnz"] Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.624405 4692 projected.go:194] Error preparing data for projected volume kube-api-access-sv2qt for pod openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.624469 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-kube-api-access-sv2qt podName:37e678a0-8c3a-4ff9-9203-9543ddd11bbb nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.124450864 +0000 UTC m=+124.949186445 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-sv2qt" (UniqueName: "kubernetes.io/projected/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-kube-api-access-sv2qt") pod "openshift-config-operator-7777fb866f-5bkn6" (UID: "37e678a0-8c3a-4ff9-9203-9543ddd11bbb") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.631584 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.639280 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.642611 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.642704 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.142690252 +0000 UTC m=+124.967425833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.642773 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.643138 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.143126405 +0000 UTC m=+124.967861986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.654531 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.657391 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.672870 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.682935 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.713685 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.713924 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.726119 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.726473 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.734838 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.736096 4692 projected.go:194] Error preparing data for projected volume kube-api-access-d24t6 for pod openshift-console-operator/console-operator-58897d9998-zd4xm: failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.736206 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-kube-api-access-d24t6 podName:7b9088be-ec8e-43ad-9b73-ae106d9afdc0 nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.236178741 +0000 UTC m=+125.060914322 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-d24t6" (UniqueName: "kubernetes.io/projected/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-kube-api-access-d24t6") pod "console-operator-58897d9998-zd4xm" (UID: "7b9088be-ec8e-43ad-9b73-ae106d9afdc0") : failed to sync configmap cache: timed out waiting for the condition Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.744213 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.744762 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.244715222 +0000 UTC m=+125.069450803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.755590 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.767753 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.767849 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.780126 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.780426 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.780525 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.780601 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.783785 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.792417 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.794002 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" event={"ID":"5537a456-f771-4a55-b811-f20b50cd6446","Type":"ContainerStarted","Data":"6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.794039 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" event={"ID":"5537a456-f771-4a55-b811-f20b50cd6446","Type":"ContainerStarted","Data":"0f688c61781d157cd9f4d3521383f882b2391404f59f65310fab48206e98737b"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.794498 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.800021 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" event={"ID":"553737d8-1213-41ef-a0ee-d1f85d749634","Type":"ContainerStarted","Data":"96bca5c6d340bb79f7c2de223a2161a8ecafbcb4e25f2bfa010107032ef06696"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.804111 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" event={"ID":"deb95cf2-78dc-43a2-acce-73f8be8e9f03","Type":"ContainerStarted","Data":"7f950156b79376f4ece47c37121aa07182f6ad7f78a05c0e4222fdf0229f9ba0"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.804168 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" event={"ID":"deb95cf2-78dc-43a2-acce-73f8be8e9f03","Type":"ContainerStarted","Data":"173a82f3ba571b0d7ac1fa29573710d0549ef200f584a43ab83992e57bda9c20"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.807899 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" event={"ID":"07c9ac14-6e46-4243-a490-6467522ae8e4","Type":"ContainerStarted","Data":"4ba3ea8b55f148eb7ff62c425fb7e764e8760cca2909716741cb925a9dcd6eb4"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.807929 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" event={"ID":"07c9ac14-6e46-4243-a490-6467522ae8e4","Type":"ContainerStarted","Data":"66b4dff37343ca63bef4851dccb3e3be0ad9006c41f981b43222180c64b05847"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.812455 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.816306 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.822393 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbw5p\" (UniqueName: \"kubernetes.io/projected/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-kube-api-access-sbw5p\") pod \"marketplace-operator-79b997595-rdvhn\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.822444 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-d5xx5" event={"ID":"cb17c405-b001-4e0a-8721-776ce83db499","Type":"ContainerStarted","Data":"6f3c65aef3672c1131dbd6a2ce59d65fd73a58256e41022ae8c1bf20084557d3"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.822504 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-d5xx5" event={"ID":"cb17c405-b001-4e0a-8721-776ce83db499","Type":"ContainerStarted","Data":"ce2f56d4e93c8125d798cd5295de12529b678bc1ec0982f45d85f6d009caac5e"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.823358 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550802-w82g5" event={"ID":"64c0a3f4-29f0-42fa-83bc-83e15171ed77","Type":"ContainerStarted","Data":"865ca57475140c4720d8fa592829144739c7092a224cf9b0beff03dfa6ba0035"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.824477 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jsgfm" event={"ID":"7880dcf7-b6e3-4620-ba8b-013f207b02e0","Type":"ContainerStarted","Data":"e4d18a73cc0d0c75a8bf38c0ca1d2c72dc96e07f2495133b101a8a061d970054"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.824519 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jsgfm" event={"ID":"7880dcf7-b6e3-4620-ba8b-013f207b02e0","Type":"ContainerStarted","Data":"5c271df51e65388ca19ebb1f1317ceaf823b9771e2811a16c67e663827a2bc30"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.826812 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vbp6d" event={"ID":"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5","Type":"ContainerStarted","Data":"66f070e5da88f68e858b75bf454fb34c2f43896df1688b049c721cd335f6d27d"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.826841 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vbp6d" event={"ID":"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5","Type":"ContainerStarted","Data":"cd0fec2bc09f98f7c72a467823d13297a568ec82d0b1260cee6716fa3370fb45"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.829763 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" event={"ID":"d4063de5-2cdb-4d86-a74f-c98af0aae38d","Type":"ContainerStarted","Data":"2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.829797 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" event={"ID":"d4063de5-2cdb-4d86-a74f-c98af0aae38d","Type":"ContainerStarted","Data":"a9341bbcca19f0316c60526aa64c3e5761de4566f421911a4e09c5249e1143f3"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.830126 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.831404 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fct6h" event={"ID":"fbcc5285-0190-40e3-968b-2ff8de9bb6a9","Type":"ContainerStarted","Data":"c278d5e912bfba5703138c5c6e6ef0f0f59d4cf3f189b6a915937365ea16b652"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.831462 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fct6h" event={"ID":"fbcc5285-0190-40e3-968b-2ff8de9bb6a9","Type":"ContainerStarted","Data":"9415b6aab3334c63d62d859f6d6cada965c1343bae69abe616246c99079dbfe4"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.833002 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" event={"ID":"f93524c6-53a2-4df0-873b-241f5ff69a9d","Type":"ContainerStarted","Data":"122ba6559fb27f8d63d68a7362baacadac42c138c9b187c2bc1d446824a95aa5"} Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.847076 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.849336 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.349314591 +0000 UTC m=+125.174050172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.858189 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.859322 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.911811 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.917864 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.948428 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.948789 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.448763513 +0000 UTC m=+125.273499094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.949019 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.949259 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nzq9\" (UniqueName: \"kubernetes.io/projected/ddda8795-d770-4f73-a1c2-c988092f7fa4-kube-api-access-8nzq9\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:03 crc kubenswrapper[4692]: E0309 09:22:03.949409 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.449393052 +0000 UTC m=+125.274128823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.953374 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nzq9\" (UniqueName: \"kubernetes.io/projected/ddda8795-d770-4f73-a1c2-c988092f7fa4-kube-api-access-8nzq9\") pod \"machine-approver-56656f9798-nvv5s\" (UID: \"ddda8795-d770-4f73-a1c2-c988092f7fa4\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.985600 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 09 09:22:03 crc kubenswrapper[4692]: I0309 09:22:03.994933 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.041119 4692 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pbl6c container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" start-of-body= Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.041223 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" podUID="5537a456-f771-4a55-b811-f20b50cd6446" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.050639 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.050897 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm5sx\" (UniqueName: \"kubernetes.io/projected/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-kube-api-access-bm5sx\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.050943 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdwrf\" (UniqueName: \"kubernetes.io/projected/4607c813-5202-460e-b583-a03dc13eee01-kube-api-access-tdwrf\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.050966 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fcz2\" (UniqueName: \"kubernetes.io/projected/c279ea73-49f4-441d-9a36-bedd8388437e-kube-api-access-4fcz2\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.051672 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.551642538 +0000 UTC m=+125.376378269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.055658 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm5sx\" (UniqueName: \"kubernetes.io/projected/1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca-kube-api-access-bm5sx\") pod \"etcd-operator-b45778765-9vg42\" (UID: \"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca\") " pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.056188 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdwrf\" (UniqueName: \"kubernetes.io/projected/4607c813-5202-460e-b583-a03dc13eee01-kube-api-access-tdwrf\") pod \"authentication-operator-69f744f599-qpk8p\" (UID: \"4607c813-5202-460e-b583-a03dc13eee01\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.056624 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fcz2\" (UniqueName: \"kubernetes.io/projected/c279ea73-49f4-441d-9a36-bedd8388437e-kube-api-access-4fcz2\") pod \"apiserver-76f77b778f-4sdrj\" (UID: \"c279ea73-49f4-441d-9a36-bedd8388437e\") " pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.126311 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.155986 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.156027 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vtnd\" (UniqueName: \"kubernetes.io/projected/146e15b3-9b76-4a85-acb5-85530b4fbdc2-kube-api-access-6vtnd\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.156076 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv2qt\" (UniqueName: \"kubernetes.io/projected/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-kube-api-access-sv2qt\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.156480 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.656465944 +0000 UTC m=+125.481201525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.159856 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv2qt\" (UniqueName: \"kubernetes.io/projected/37e678a0-8c3a-4ff9-9203-9543ddd11bbb-kube-api-access-sv2qt\") pod \"openshift-config-operator-7777fb866f-5bkn6\" (UID: \"37e678a0-8c3a-4ff9-9203-9543ddd11bbb\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.164650 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vtnd\" (UniqueName: \"kubernetes.io/projected/146e15b3-9b76-4a85-acb5-85530b4fbdc2-kube-api-access-6vtnd\") pod \"openshift-controller-manager-operator-756b6f6bc6-qjf2f\" (UID: \"146e15b3-9b76-4a85-acb5-85530b4fbdc2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.174475 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.174525 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.195569 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.205365 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.235547 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.240395 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.256930 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.257238 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d24t6\" (UniqueName: \"kubernetes.io/projected/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-kube-api-access-d24t6\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.258474 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.758443403 +0000 UTC m=+125.583179034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.265992 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d24t6\" (UniqueName: \"kubernetes.io/projected/7b9088be-ec8e-43ad-9b73-ae106d9afdc0-kube-api-access-d24t6\") pod \"console-operator-58897d9998-zd4xm\" (UID: \"7b9088be-ec8e-43ad-9b73-ae106d9afdc0\") " pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.271831 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7gw55" podStartSLOduration=59.271065979 podStartE2EDuration="59.271065979s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:04.205988789 +0000 UTC m=+125.030724380" watchObservedRunningTime="2026-03-09 09:22:04.271065979 +0000 UTC m=+125.095801560" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.295616 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j"] Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.298727 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.302428 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.335875 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.340575 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.364999 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.365441 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.865425865 +0000 UTC m=+125.690161446 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.373046 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.379119 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.445668 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" podStartSLOduration=59.445649519 podStartE2EDuration="59.445649519s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:04.40875567 +0000 UTC m=+125.233491251" watchObservedRunningTime="2026-03-09 09:22:04.445649519 +0000 UTC m=+125.270385100" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.466527 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.466725 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.966699982 +0000 UTC m=+125.791435563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.466842 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.467240 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:04.967230168 +0000 UTC m=+125.791965969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.566404 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.582964 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.586249 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.587700 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.087677592 +0000 UTC m=+125.912413173 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.587836 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.589054 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.089042864 +0000 UTC m=+125.913778445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.663563 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.663656 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.674510 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.688770 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.688956 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.188931989 +0000 UTC m=+126.013667570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.689293 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.689693 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.189685012 +0000 UTC m=+126.014420593 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.730436 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv"] Mar 09 09:22:04 crc kubenswrapper[4692]: W0309 09:22:04.753489 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf96b54e5_804b_4d9e_9df9_538352c52073.slice/crio-5307770c0427ac67f58c5e90e32674d8e3abf5cbe12c7c389ef27085cfa41d13 WatchSource:0}: Error finding container 5307770c0427ac67f58c5e90e32674d8e3abf5cbe12c7c389ef27085cfa41d13: Status 404 returned error can't find the container with id 5307770c0427ac67f58c5e90e32674d8e3abf5cbe12c7c389ef27085cfa41d13 Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.794505 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.794723 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9dn4z"] Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.794914 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.29488638 +0000 UTC m=+126.119622151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.795029 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.795785 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.295770637 +0000 UTC m=+126.120506208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.856923 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" event={"ID":"f93524c6-53a2-4df0-873b-241f5ff69a9d","Type":"ContainerStarted","Data":"3284cf02e7655497ef6a5b57ac1dc563970f8dce7f859e77599232ebfb218df0"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.860405 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" event={"ID":"f96b54e5-804b-4d9e-9df9-538352c52073","Type":"ContainerStarted","Data":"5307770c0427ac67f58c5e90e32674d8e3abf5cbe12c7c389ef27085cfa41d13"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.866006 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794"] Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.871454 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-snq8b"] Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.887402 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vbp6d" event={"ID":"51b39b63-d3ad-4a29-b7b5-6cede45e2fa5","Type":"ContainerStarted","Data":"0a564ede9f0cafe9b8a101437542a8d870cd0cfa16baebaddd6d20aff74e02ce"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.887670 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.894317 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rfkdj"] Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.895565 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" event={"ID":"03b56c7f-0613-419a-8a32-fe678a4a3e93","Type":"ContainerStarted","Data":"e6d737ab0cea22fae3974e38319e9a266ba31bae367fd4f6c372821a73351ce5"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.898200 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.898336 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.398315303 +0000 UTC m=+126.223050874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.898844 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:04 crc kubenswrapper[4692]: E0309 09:22:04.899764 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.399747757 +0000 UTC m=+126.224483338 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.901457 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" event={"ID":"ddda8795-d770-4f73-a1c2-c988092f7fa4","Type":"ContainerStarted","Data":"d7e641f89b63284b483af6f3e8c2788fea3d97cd61b7a5829ef562ef1e33ed4d"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.903943 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9dn4z" event={"ID":"68ff6f00-0ac5-41d9-b762-933944bdb507","Type":"ContainerStarted","Data":"1c39cc2cc29fe2164d849d32b087d0c8af5e739b56f96cb512f711dd38d06fc9"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.906002 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" event={"ID":"553737d8-1213-41ef-a0ee-d1f85d749634","Type":"ContainerStarted","Data":"3623d5ae270f47bd8f69fc6ac7971857619fa36897b2ab3390e76a37e3a111da"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.908953 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" event={"ID":"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334","Type":"ContainerStarted","Data":"44778371378d013f077f3043f653cd00c46e674cae60325ccc8a7a921f01f970"} Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.918598 4692 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-pbl6c container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" start-of-body= Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.918667 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" podUID="5537a456-f771-4a55-b811-f20b50cd6446" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" Mar 09 09:22:04 crc kubenswrapper[4692]: W0309 09:22:04.926102 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7e9d340_f169_49ff_82ae_21089bdb28f6.slice/crio-36f26c93559f7c53dd3632e6f150d339d502d63ae15d8db5735c71033fde7184 WatchSource:0}: Error finding container 36f26c93559f7c53dd3632e6f150d339d502d63ae15d8db5735c71033fde7184: Status 404 returned error can't find the container with id 36f26c93559f7c53dd3632e6f150d339d502d63ae15d8db5735c71033fde7184 Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.926392 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.947278 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:04 crc kubenswrapper[4692]: I0309 09:22:04.999830 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.000058 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.500017323 +0000 UTC m=+126.324752924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.000690 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.004994 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.504972165 +0000 UTC m=+126.329707746 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.100666 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.106416 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.106786 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.606735017 +0000 UTC m=+126.431470608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.126278 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.126379 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.141800 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" podStartSLOduration=60.141778849 podStartE2EDuration="1m0.141778849s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:05.141059597 +0000 UTC m=+125.965795188" watchObservedRunningTime="2026-03-09 09:22:05.141778849 +0000 UTC m=+125.966514430" Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.209044 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.209474 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.709454809 +0000 UTC m=+126.534190580 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.278842 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79"] Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.283203 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod981360f1_a9de_4035_bbb0_2fae838061f5.slice/crio-f6fdeaaf9b872265f73667eea4c11608c01b4b9d95191ea88bb07c6af24bcf3e WatchSource:0}: Error finding container f6fdeaaf9b872265f73667eea4c11608c01b4b9d95191ea88bb07c6af24bcf3e: Status 404 returned error can't find the container with id f6fdeaaf9b872265f73667eea4c11608c01b4b9d95191ea88bb07c6af24bcf3e Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.314420 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.314530 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.814502872 +0000 UTC m=+126.639238453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.314865 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.315256 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.815245154 +0000 UTC m=+126.639980735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.317003 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.328241 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.337569 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.339225 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-4sdrj"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.347430 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.357769 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.360326 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.360867 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.363262 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.364605 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.368507 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdvhn"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.370073 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.416156 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.416348 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.916305315 +0000 UTC m=+126.741040896 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.416537 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.416998 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:05.916987366 +0000 UTC m=+126.741722947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.517355 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.517598 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.017576332 +0000 UTC m=+126.842311913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.517719 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.518146 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.018132479 +0000 UTC m=+126.842868060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.619506 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.619722 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.119679365 +0000 UTC m=+126.944414946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.619917 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.620335 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.120325135 +0000 UTC m=+126.945060716 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.721426 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.721614 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.221595342 +0000 UTC m=+127.046330923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.721722 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.722095 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.222076427 +0000 UTC m=+127.046812078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.783731 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8hxvb" podStartSLOduration=60.783712362 podStartE2EDuration="1m0.783712362s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:05.781612908 +0000 UTC m=+126.606348489" watchObservedRunningTime="2026-03-09 09:22:05.783712362 +0000 UTC m=+126.608447943" Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.823407 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.823934 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.323914222 +0000 UTC m=+127.148649803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.871872 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vbp6d" podStartSLOduration=7.871850148 podStartE2EDuration="7.871850148s" podCreationTimestamp="2026-03-09 09:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:05.871523368 +0000 UTC m=+126.696258949" watchObservedRunningTime="2026-03-09 09:22:05.871850148 +0000 UTC m=+126.696585749" Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.891356 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-9vg42"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.896896 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qpk8p"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.899041 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-rddqb"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.903941 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-vdjpl"] Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.909910 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod385a2958_770a_4ce1_9b88_b897cb08c3eb.slice/crio-a91bccafba75da7caedc7ba1d3b5a937887542bb9baa57986f15c7c55f2b87ff WatchSource:0}: Error finding container a91bccafba75da7caedc7ba1d3b5a937887542bb9baa57986f15c7c55f2b87ff: Status 404 returned error can't find the container with id a91bccafba75da7caedc7ba1d3b5a937887542bb9baa57986f15c7c55f2b87ff Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.913732 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod700952b9_c6cb_403f_855b_b3cbd855e845.slice/crio-0f3b84abd1c4008b9e606b30fbe7c155a41e8b73ecf8041b5be5cd861e2ebec2 WatchSource:0}: Error finding container 0f3b84abd1c4008b9e606b30fbe7c155a41e8b73ecf8041b5be5cd861e2ebec2: Status 404 returned error can't find the container with id 0f3b84abd1c4008b9e606b30fbe7c155a41e8b73ecf8041b5be5cd861e2ebec2 Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.914450 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf73e489b_1a01_4608_a493_e0ce7dd2c2c5.slice/crio-2ff8ed88c8eda786b4a27d734242b4a16b4ab7a0fad1cc649f4d60330c820ab5 WatchSource:0}: Error finding container 2ff8ed88c8eda786b4a27d734242b4a16b4ab7a0fad1cc649f4d60330c820ab5: Status 404 returned error can't find the container with id 2ff8ed88c8eda786b4a27d734242b4a16b4ab7a0fad1cc649f4d60330c820ab5 Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.917150 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6"] Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.917737 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64cbcb6e_701d_45c4_8293_945716329cef.slice/crio-13c88420e63ae0df73f9ebc4fa07decac8406f0df45864bdce673c8929b6803f WatchSource:0}: Error finding container 13c88420e63ae0df73f9ebc4fa07decac8406f0df45864bdce673c8929b6803f: Status 404 returned error can't find the container with id 13c88420e63ae0df73f9ebc4fa07decac8406f0df45864bdce673c8929b6803f Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.920555 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod543c31b5_3ba9_4e04_aeaf_3ba7315da541.slice/crio-c0a93ac459ac68a7f98facec7304a6591c66806e16084e4617f275196b5212c9 WatchSource:0}: Error finding container c0a93ac459ac68a7f98facec7304a6591c66806e16084e4617f275196b5212c9: Status 404 returned error can't find the container with id c0a93ac459ac68a7f98facec7304a6591c66806e16084e4617f275196b5212c9 Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.921527 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zd4xm"] Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.924753 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:05 crc kubenswrapper[4692]: E0309 09:22:05.925126 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.425111526 +0000 UTC m=+127.249847107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.931840 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12f530da_08e1_42ee_a635_ca648ee4f5b3.slice/crio-af189d118e1652d6e9277d0b56831e4a6db290b80c6e0c2f3f4f47cf7df5ab97 WatchSource:0}: Error finding container af189d118e1652d6e9277d0b56831e4a6db290b80c6e0c2f3f4f47cf7df5ab97: Status 404 returned error can't find the container with id af189d118e1652d6e9277d0b56831e4a6db290b80c6e0c2f3f4f47cf7df5ab97 Mar 09 09:22:05 crc kubenswrapper[4692]: W0309 09:22:05.979668 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96013f1f_d2b0_42ec_a4f3_3626f48be8e1.slice/crio-f16e39be91f32263c836c1e1974c4418ec7eb12ffb0525ae6c9188746542c632 WatchSource:0}: Error finding container f16e39be91f32263c836c1e1974c4418ec7eb12ffb0525ae6c9188746542c632: Status 404 returned error can't find the container with id f16e39be91f32263c836c1e1974c4418ec7eb12ffb0525ae6c9188746542c632 Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.995708 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" event={"ID":"d59db20e-6720-4cc9-bfd8-b3477c7298cf","Type":"ContainerStarted","Data":"ddb08cea9ac7dba0ee172e5329771e95503b14ce4fa6614d836d14d235e9596d"} Mar 09 09:22:05 crc kubenswrapper[4692]: I0309 09:22:05.997598 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" event={"ID":"c279ea73-49f4-441d-9a36-bedd8388437e","Type":"ContainerStarted","Data":"f6b98237b5edb569a0cb57e064ea32b031367dd8a1c6d3995f549719193705d9"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.001215 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" event={"ID":"146e15b3-9b76-4a85-acb5-85530b4fbdc2","Type":"ContainerStarted","Data":"7d1d2e1c4d376822e4b563adba3fae257b63c0dc0af8c6dd616e690b28ae53be"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.002367 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" event={"ID":"8385d993-0021-4066-a9c6-57bf7ee4be33","Type":"ContainerStarted","Data":"da407dbe508faeff4bf4e36a8527ef8c3d0b04ea6e2ca6d5325acbcf4f9fc5d6"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.006755 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" event={"ID":"9c16e62e-9744-45c8-88d9-0c289df743a1","Type":"ContainerStarted","Data":"e41dd39c8335c7172f3a2e607b2a6c57afb8c2a82d9fcc1fe4e41e704c05db3a"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.012869 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" event={"ID":"981360f1-a9de-4035-bbb0-2fae838061f5","Type":"ContainerStarted","Data":"f6fdeaaf9b872265f73667eea4c11608c01b4b9d95191ea88bb07c6af24bcf3e"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.025832 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:06 crc kubenswrapper[4692]: W0309 09:22:06.028072 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93ccefab_c47d_4725_a9ff_55cc50f678f0.slice/crio-2338795ff00e2ef9bc1b9806b94aa3d61ae5a45b99a9398b33d86f8aadf0ea02 WatchSource:0}: Error finding container 2338795ff00e2ef9bc1b9806b94aa3d61ae5a45b99a9398b33d86f8aadf0ea02: Status 404 returned error can't find the container with id 2338795ff00e2ef9bc1b9806b94aa3d61ae5a45b99a9398b33d86f8aadf0ea02 Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.028392 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.528357134 +0000 UTC m=+127.353092715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.028842 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-snq8b" event={"ID":"aa908134-de3c-43e0-ace1-94fc548861c8","Type":"ContainerStarted","Data":"dc9c2d2c61f146e814c868ed81c6df6a9b7141dd1a29db4bb806150215e101bd"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.034451 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-bhnwz" podStartSLOduration=61.03443292 podStartE2EDuration="1m1.03443292s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.03377868 +0000 UTC m=+126.858514261" watchObservedRunningTime="2026-03-09 09:22:06.03443292 +0000 UTC m=+126.859168501" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.053385 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.057197 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.557179656 +0000 UTC m=+127.381915247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.068913 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" event={"ID":"385a2958-770a-4ce1-9b88-b897cb08c3eb","Type":"ContainerStarted","Data":"a91bccafba75da7caedc7ba1d3b5a937887542bb9baa57986f15c7c55f2b87ff"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.116453 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" event={"ID":"974389aa-40cb-4191-a85b-d14e1aa175af","Type":"ContainerStarted","Data":"543069763f12aa1267c5f52aee940f10f99b51cabae1f6d9a4e9fc7b4d43669e"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.120297 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" event={"ID":"d7e9d340-f169-49ff-82ae-21089bdb28f6","Type":"ContainerStarted","Data":"36f26c93559f7c53dd3632e6f150d339d502d63ae15d8db5735c71033fde7184"} Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.126600 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:06 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:06 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:06 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.126641 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.158330 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.158625 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.658611348 +0000 UTC m=+127.483346929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.183348 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" podStartSLOduration=8.183330074 podStartE2EDuration="8.183330074s" podCreationTimestamp="2026-03-09 09:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.182364314 +0000 UTC m=+127.007099905" watchObservedRunningTime="2026-03-09 09:22:06.183330074 +0000 UTC m=+127.008065675" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.220258 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jsgfm" podStartSLOduration=8.220237623 podStartE2EDuration="8.220237623s" podCreationTimestamp="2026-03-09 09:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.219134869 +0000 UTC m=+127.043870460" watchObservedRunningTime="2026-03-09 09:22:06.220237623 +0000 UTC m=+127.044973204" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.263335 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.263721 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:06.763702922 +0000 UTC m=+127.588438713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.502558 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-ldfbc" podStartSLOduration=61.502524446 podStartE2EDuration="1m1.502524446s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.464666048 +0000 UTC m=+127.289401629" watchObservedRunningTime="2026-03-09 09:22:06.502524446 +0000 UTC m=+127.327260027" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.503101 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-4qpm7" podStartSLOduration=61.503095854 podStartE2EDuration="1m1.503095854s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.499899736 +0000 UTC m=+127.324635327" watchObservedRunningTime="2026-03-09 09:22:06.503095854 +0000 UTC m=+127.327831435" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.540616 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rxzs8" podStartSLOduration=61.540599651 podStartE2EDuration="1m1.540599651s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.53830115 +0000 UTC m=+127.363036731" watchObservedRunningTime="2026-03-09 09:22:06.540599651 +0000 UTC m=+127.365335232" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.580854 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.581517 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.081490531 +0000 UTC m=+127.906226122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.598384 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fct6h" podStartSLOduration=8.598363347 podStartE2EDuration="8.598363347s" podCreationTimestamp="2026-03-09 09:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.595226652 +0000 UTC m=+127.419962243" watchObservedRunningTime="2026-03-09 09:22:06.598363347 +0000 UTC m=+127.423098918" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.681297 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" podStartSLOduration=62.681232112000004 podStartE2EDuration="1m2.681232112s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.680435078 +0000 UTC m=+127.505170659" watchObservedRunningTime="2026-03-09 09:22:06.681232112 +0000 UTC m=+127.505967703" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.682283 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.682989 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.182971685 +0000 UTC m=+128.007707266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.783598 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.783855 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.28383291 +0000 UTC m=+128.108568501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.784333 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.784707 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.284696736 +0000 UTC m=+128.109432327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.885246 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.885736 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.385716326 +0000 UTC m=+128.210451917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.921356 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-d5xx5" podStartSLOduration=61.921330945 podStartE2EDuration="1m1.921330945s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:06.914590179 +0000 UTC m=+127.739325780" watchObservedRunningTime="2026-03-09 09:22:06.921330945 +0000 UTC m=+127.746066526" Mar 09 09:22:06 crc kubenswrapper[4692]: I0309 09:22:06.987365 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:06 crc kubenswrapper[4692]: E0309 09:22:06.987693 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.487680544 +0000 UTC m=+128.312416125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.029122 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-c9rx4"] Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.088432 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.088810 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.588793367 +0000 UTC m=+128.413528948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.128971 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:07 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:07 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:07 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.129029 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.140944 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rddqb" event={"ID":"7b8b20b2-f067-4abd-adc3-7b1a2bf36742","Type":"ContainerStarted","Data":"5cad8fe74b1f7e20fb38ebb95970488fbab7029de6987687ea402c3a84250bdf"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.144048 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" event={"ID":"7b9088be-ec8e-43ad-9b73-ae106d9afdc0","Type":"ContainerStarted","Data":"84f880b265b58a976ddcd58d5a697abd84d7b2ac8524432cf69f83a682e17eb7"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.145238 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" event={"ID":"d59db20e-6720-4cc9-bfd8-b3477c7298cf","Type":"ContainerStarted","Data":"1adf46dfc15dd6bc710ef1e51f1b2f8dccc616e63f2207b698bc6f57f4615c8b"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.145945 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" event={"ID":"df306321-4c89-4909-834c-48ae131ac196","Type":"ContainerStarted","Data":"a09a0b76566cede0b1bb08e7f68f83be416095c678b0ea91c9a334f4e4e45517"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.147697 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" event={"ID":"8385d993-0021-4066-a9c6-57bf7ee4be33","Type":"ContainerStarted","Data":"ac9ea6570180d5b1e2a06faac6634bcaf2ceba6357decd021f5909945c1b6efd"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.148720 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" event={"ID":"ddda8795-d770-4f73-a1c2-c988092f7fa4","Type":"ContainerStarted","Data":"75d9c112bf3ed70c00a0da6597ee173044f1964c62c7c881491d0fcfe0f31abc"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.149653 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9dn4z" event={"ID":"68ff6f00-0ac5-41d9-b762-933944bdb507","Type":"ContainerStarted","Data":"82ce78e70af66349f259066fa93907c7787ad8404ea652e1e2ebd585e09d7b0b"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.154489 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" event={"ID":"700952b9-c6cb-403f-855b-b3cbd855e845","Type":"ContainerStarted","Data":"0f3b84abd1c4008b9e606b30fbe7c155a41e8b73ecf8041b5be5cd861e2ebec2"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.155434 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" event={"ID":"37e678a0-8c3a-4ff9-9203-9543ddd11bbb","Type":"ContainerStarted","Data":"3fe7b7fbb26faf6d9a7201ba4a16176821fa66bb873948e578b31dbe3d97d179"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.156459 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" event={"ID":"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca","Type":"ContainerStarted","Data":"8df39ef8eaf9894bbeb06f5203d97b64d46a8938de73d0211609bbae3e741a60"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.159429 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" event={"ID":"f96b54e5-804b-4d9e-9df9-538352c52073","Type":"ContainerStarted","Data":"ac247f9bfc4074d231deb121d854bad9b6a80025df752a0138fad95d1a0dbfb0"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.160787 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" event={"ID":"981360f1-a9de-4035-bbb0-2fae838061f5","Type":"ContainerStarted","Data":"860d3c3b1ee54449a50e0e53790b0e6d44a3201c34ea5bf1a136d4080ea1624b"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.161326 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" event={"ID":"12f530da-08e1-42ee-a635-ca648ee4f5b3","Type":"ContainerStarted","Data":"af189d118e1652d6e9277d0b56831e4a6db290b80c6e0c2f3f4f47cf7df5ab97"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.162135 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" event={"ID":"543c31b5-3ba9-4e04-aeaf-3ba7315da541","Type":"ContainerStarted","Data":"c0a93ac459ac68a7f98facec7304a6591c66806e16084e4617f275196b5212c9"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.163240 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" event={"ID":"553737d8-1213-41ef-a0ee-d1f85d749634","Type":"ContainerStarted","Data":"abbffeb5faa83bfb67c5c2da26cdf51fdfcb668d6b3a18b8ca0df565e9f79075"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.164461 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" event={"ID":"14a9e924-9e6f-4a5e-82f4-ea2b52ae4334","Type":"ContainerStarted","Data":"ae5fcb4cf49a579b3eb9d322bcb87120c7b50ba0a8504ba1f5ef07d0aacd9811"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.165872 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" event={"ID":"96013f1f-d2b0-42ec-a4f3-3626f48be8e1","Type":"ContainerStarted","Data":"f16e39be91f32263c836c1e1974c4418ec7eb12ffb0525ae6c9188746542c632"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.166990 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" event={"ID":"93ccefab-c47d-4725-a9ff-55cc50f678f0","Type":"ContainerStarted","Data":"2338795ff00e2ef9bc1b9806b94aa3d61ae5a45b99a9398b33d86f8aadf0ea02"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.167998 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" event={"ID":"f73e489b-1a01-4608-a493-e0ce7dd2c2c5","Type":"ContainerStarted","Data":"2ff8ed88c8eda786b4a27d734242b4a16b4ab7a0fad1cc649f4d60330c820ab5"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.169244 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" event={"ID":"d7e9d340-f169-49ff-82ae-21089bdb28f6","Type":"ContainerStarted","Data":"cc4da9990ccba6b4b9e686a3e32c8fdac1d444a0c57ceb573954310d70222d72"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.170058 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-snq8b" event={"ID":"aa908134-de3c-43e0-ace1-94fc548861c8","Type":"ContainerStarted","Data":"70a298d5a5e71f3b1a4df0cda89d4cc48ebf01866d50aa809c2bd2d22597bbf5"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.170706 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-snq8b" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.171412 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" event={"ID":"64cbcb6e-701d-45c4-8293-945716329cef","Type":"ContainerStarted","Data":"13c88420e63ae0df73f9ebc4fa07decac8406f0df45864bdce673c8929b6803f"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.171820 4692 patch_prober.go:28] interesting pod/downloads-7954f5f757-snq8b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.171858 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-snq8b" podUID="aa908134-de3c-43e0-ace1-94fc548861c8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.174611 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" event={"ID":"146e15b3-9b76-4a85-acb5-85530b4fbdc2","Type":"ContainerStarted","Data":"01ae47b7556fd080878d5e718fdcfeb2f94298679d364367b6a9515ef0522533"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.175075 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" podStartSLOduration=63.175054595 podStartE2EDuration="1m3.175054595s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:07.174363164 +0000 UTC m=+127.999098755" watchObservedRunningTime="2026-03-09 09:22:07.175054595 +0000 UTC m=+127.999790176" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.176275 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" event={"ID":"4607c813-5202-460e-b583-a03dc13eee01","Type":"ContainerStarted","Data":"05cb57a82e4c5873a419fc71b7b7af49acf3eff574bdea2fe8d5b3295910dc98"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.177557 4692 generic.go:334] "Generic (PLEG): container finished" podID="c279ea73-49f4-441d-9a36-bedd8388437e" containerID="0d904157ea681cc1ff23bb5d85dde646f2f9a975b6286a70eef772d7e8721681" exitCode=0 Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.177710 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" gracePeriod=30 Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.178316 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" event={"ID":"c279ea73-49f4-441d-9a36-bedd8388437e","Type":"ContainerDied","Data":"0d904157ea681cc1ff23bb5d85dde646f2f9a975b6286a70eef772d7e8721681"} Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.189865 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bng8j" podStartSLOduration=62.189849798 podStartE2EDuration="1m2.189849798s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:07.188676342 +0000 UTC m=+128.013411933" watchObservedRunningTime="2026-03-09 09:22:07.189849798 +0000 UTC m=+128.014585379" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.192820 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.193141 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.693131168 +0000 UTC m=+128.517866749 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.208957 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-tr7zk" podStartSLOduration=62.208941071 podStartE2EDuration="1m2.208941071s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:07.207570039 +0000 UTC m=+128.032305640" watchObservedRunningTime="2026-03-09 09:22:07.208941071 +0000 UTC m=+128.033676652" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.223468 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-snq8b" podStartSLOduration=62.223446965 podStartE2EDuration="1m2.223446965s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:07.222689142 +0000 UTC m=+128.047424723" watchObservedRunningTime="2026-03-09 09:22:07.223446965 +0000 UTC m=+128.048182546" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.254807 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-l8794" podStartSLOduration=63.254784524 podStartE2EDuration="1m3.254784524s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:07.253591457 +0000 UTC m=+128.078327048" watchObservedRunningTime="2026-03-09 09:22:07.254784524 +0000 UTC m=+128.079520105" Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.294028 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.294441 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.794422406 +0000 UTC m=+128.619157997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.294475 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.295507 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.795488128 +0000 UTC m=+128.620223869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.396543 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.396945 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:07.896931181 +0000 UTC m=+128.721666762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.497486 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.500953 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.000936962 +0000 UTC m=+128.825672543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.600313 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.602223 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.102196728 +0000 UTC m=+128.926932309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.603770 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.604371 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.104359824 +0000 UTC m=+128.929095405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.710856 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.711622 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.211602714 +0000 UTC m=+129.036338295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.817646 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.817998 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.317986297 +0000 UTC m=+129.142721878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:07 crc kubenswrapper[4692]: I0309 09:22:07.920702 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:07 crc kubenswrapper[4692]: E0309 09:22:07.921525 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.421494503 +0000 UTC m=+129.246230144 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.022531 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.029116 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.529089644 +0000 UTC m=+129.353825225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.123407 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.123753 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.623736739 +0000 UTC m=+129.448472320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.128581 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:08 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:08 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:08 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.128642 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.191832 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" event={"ID":"37e678a0-8c3a-4ff9-9203-9543ddd11bbb","Type":"ContainerStarted","Data":"4af810a4f8dc46992a8ebd2da10e16c78743ffcdc5a059bbd1921d074fc86eef"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.212726 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" event={"ID":"64cbcb6e-701d-45c4-8293-945716329cef","Type":"ContainerStarted","Data":"02c46a982c5e903122e11626087665f0afd1aa1fd9c889049c55e0e96b21fcf8"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.225856 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.226184 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.726148081 +0000 UTC m=+129.550883662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.233502 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" event={"ID":"9c16e62e-9744-45c8-88d9-0c289df743a1","Type":"ContainerStarted","Data":"ab70e7cb8f838eef1760fd1a6d4f8c86d6ed381583310145a313255f3570a71c"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.235101 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.237899 4692 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-29jhq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.237944 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" podUID="9c16e62e-9744-45c8-88d9-0c289df743a1" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.240219 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-mpwzq" podStartSLOduration=63.240201251 podStartE2EDuration="1m3.240201251s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.23820504 +0000 UTC m=+129.062940641" watchObservedRunningTime="2026-03-09 09:22:08.240201251 +0000 UTC m=+129.064936832" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.245792 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" event={"ID":"1c17d21c-66aa-43d1-9c2a-dacc31e6b8ca","Type":"ContainerStarted","Data":"e10e485247b79ba146188011cfab7f4f30682bc23296ba2b7efdf820cda0b797"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.254097 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9dn4z" event={"ID":"68ff6f00-0ac5-41d9-b762-933944bdb507","Type":"ContainerStarted","Data":"c31e2f0799c6522bdc2e094483977ff8f8de943a6b3484552a130e57cdad2cd7"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.259402 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" event={"ID":"543c31b5-3ba9-4e04-aeaf-3ba7315da541","Type":"ContainerStarted","Data":"b39daaa559b72d10a428159daae47de99b6211e5ecd96c481dacc5c3309a7cc1"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.259444 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" event={"ID":"543c31b5-3ba9-4e04-aeaf-3ba7315da541","Type":"ContainerStarted","Data":"c4e3aba034e4efbc5fa8bec1fb0dfeee47b2ecca1451ee82ec2562d3b55942d5"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.263123 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-rddqb" event={"ID":"7b8b20b2-f067-4abd-adc3-7b1a2bf36742","Type":"ContainerStarted","Data":"c84759c1b1051bf60d00efc90e7b278d29cbcac60c32acebabc7ce19955668a9"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.263868 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" podStartSLOduration=63.263851844 podStartE2EDuration="1m3.263851844s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.263674329 +0000 UTC m=+129.088409910" watchObservedRunningTime="2026-03-09 09:22:08.263851844 +0000 UTC m=+129.088587425" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.268472 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" event={"ID":"974389aa-40cb-4191-a85b-d14e1aa175af","Type":"ContainerStarted","Data":"7888e55f1777e3f7fe6084c9167746cf7da101af47891a621c0cb14d0bb8dfff"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.268529 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" event={"ID":"974389aa-40cb-4191-a85b-d14e1aa175af","Type":"ContainerStarted","Data":"07b0bf19396561b75df3227b677431e1869d08853af678cd9a9141e66360e0ec"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.270750 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" event={"ID":"12f530da-08e1-42ee-a635-ca648ee4f5b3","Type":"ContainerStarted","Data":"63cb55e8fdca437f739cbb316a0df73aca84ab69daeb865a80a6d7a822a2aa2f"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.274380 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" event={"ID":"ddda8795-d770-4f73-a1c2-c988092f7fa4","Type":"ContainerStarted","Data":"8295a0acbd79b11ae9cfe557639a010a9392848f69737241d938062249ce279e"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.277454 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" event={"ID":"93ccefab-c47d-4725-a9ff-55cc50f678f0","Type":"ContainerStarted","Data":"46b28b153ea49f1f81d143f7ca2ca7342a6606fba9b3cd1c94e5846d994505de"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.279367 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" event={"ID":"700952b9-c6cb-403f-855b-b3cbd855e845","Type":"ContainerStarted","Data":"9581348a0b9a50e8634e089968125e7b9e0b70fe06771a73f7e0e296383abd92"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.280054 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.281287 4692 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kn54j container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.281345 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" podUID="700952b9-c6cb-403f-855b-b3cbd855e845" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.281862 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" event={"ID":"7b9088be-ec8e-43ad-9b73-ae106d9afdc0","Type":"ContainerStarted","Data":"24965d749f3af27d7b415911f8d21b4b5592d1a93bed00f4343ea4649d621553"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.282007 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.283784 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" event={"ID":"96013f1f-d2b0-42ec-a4f3-3626f48be8e1","Type":"ContainerStarted","Data":"d7247aed919cff304a8b476a12e389279bfab9c4696fa35749265c5d9f1f327d"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.284019 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.286649 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" event={"ID":"8385d993-0021-4066-a9c6-57bf7ee4be33","Type":"ContainerStarted","Data":"d37b9f11a1f04a5631af24e0263fc2ccba706f9d1341f5d9026d9e17133c4e8a"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.303288 4692 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rdvhn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.303567 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" event={"ID":"c279ea73-49f4-441d-9a36-bedd8388437e","Type":"ContainerStarted","Data":"37a876c47467f2fd183b151322fbbc129c9780f96a7ad635dffd08b3656965ae"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.303327 4692 patch_prober.go:28] interesting pod/console-operator-58897d9998-zd4xm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.303673 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" podUID="7b9088be-ec8e-43ad-9b73-ae106d9afdc0" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.303601 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.315889 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" event={"ID":"385a2958-770a-4ce1-9b88-b897cb08c3eb","Type":"ContainerStarted","Data":"266950a4e41fad0e1a7a0aa6b09802639f335bf918b7df2c370635ff1eb879d2"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.315936 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" event={"ID":"385a2958-770a-4ce1-9b88-b897cb08c3eb","Type":"ContainerStarted","Data":"3f5ed507ba0a5c02b5110ec589e8eeff03a5adf53656dcbb493bc8a0fe958ea1"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.316928 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.325452 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" event={"ID":"f73e489b-1a01-4608-a493-e0ce7dd2c2c5","Type":"ContainerStarted","Data":"6059b174800f4e65da58be344f2c5f004864b1627c4513b127dc8d1cdee0777d"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.326599 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.328751 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.828732049 +0000 UTC m=+129.653467630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.330947 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9dn4z" podStartSLOduration=63.330930096 podStartE2EDuration="1m3.330930096s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.304957331 +0000 UTC m=+129.129692922" watchObservedRunningTime="2026-03-09 09:22:08.330930096 +0000 UTC m=+129.155665677" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.331230 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-9vg42" podStartSLOduration=63.331224755 podStartE2EDuration="1m3.331224755s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.328710528 +0000 UTC m=+129.153446129" watchObservedRunningTime="2026-03-09 09:22:08.331224755 +0000 UTC m=+129.155960336" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.344657 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" event={"ID":"df306321-4c89-4909-834c-48ae131ac196","Type":"ContainerStarted","Data":"26be93215b599a96527649dedc0301bd80daab0f5b42cd9e0509e506cb3caebf"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.352795 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-87j6g" podStartSLOduration=63.352775094 podStartE2EDuration="1m3.352775094s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.35197559 +0000 UTC m=+129.176711171" watchObservedRunningTime="2026-03-09 09:22:08.352775094 +0000 UTC m=+129.177510675" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.355067 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" event={"ID":"4607c813-5202-460e-b583-a03dc13eee01","Type":"ContainerStarted","Data":"6d5c040dfcd622317420518693b7a40ba9aec5d6087c47d38b225cfb55a4166c"} Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.365152 4692 patch_prober.go:28] interesting pod/downloads-7954f5f757-snq8b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.365226 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-snq8b" podUID="aa908134-de3c-43e0-ace1-94fc548861c8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.384415 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-rddqb" podStartSLOduration=63.384389191 podStartE2EDuration="1m3.384389191s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.381915125 +0000 UTC m=+129.206650726" watchObservedRunningTime="2026-03-09 09:22:08.384389191 +0000 UTC m=+129.209124792" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.400842 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7n6dg" podStartSLOduration=63.400823453 podStartE2EDuration="1m3.400823453s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.399698639 +0000 UTC m=+129.224434220" watchObservedRunningTime="2026-03-09 09:22:08.400823453 +0000 UTC m=+129.225559034" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.425538 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-nvv5s" podStartSLOduration=64.425519059 podStartE2EDuration="1m4.425519059s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.423935341 +0000 UTC m=+129.248670932" watchObservedRunningTime="2026-03-09 09:22:08.425519059 +0000 UTC m=+129.250254640" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.433268 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.436223 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:08.936209096 +0000 UTC m=+129.760944677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.482769 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" podStartSLOduration=63.482737529 podStartE2EDuration="1m3.482737529s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.447764569 +0000 UTC m=+129.272500150" watchObservedRunningTime="2026-03-09 09:22:08.482737529 +0000 UTC m=+129.307473110" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.518768 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qjf2f" podStartSLOduration=63.51874533 podStartE2EDuration="1m3.51874533s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.5184245 +0000 UTC m=+129.343160091" watchObservedRunningTime="2026-03-09 09:22:08.51874533 +0000 UTC m=+129.343480911" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.519493 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" podStartSLOduration=63.519486913 podStartE2EDuration="1m3.519486913s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.483334447 +0000 UTC m=+129.308070028" watchObservedRunningTime="2026-03-09 09:22:08.519486913 +0000 UTC m=+129.344222494" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.547636 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.548197 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.04816146 +0000 UTC m=+129.872897041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.553357 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" podStartSLOduration=63.553331228 podStartE2EDuration="1m3.553331228s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.544400985 +0000 UTC m=+129.369136576" watchObservedRunningTime="2026-03-09 09:22:08.553331228 +0000 UTC m=+129.378066819" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.572397 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" podStartSLOduration=63.572378641 podStartE2EDuration="1m3.572378641s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.570661828 +0000 UTC m=+129.395397429" watchObservedRunningTime="2026-03-09 09:22:08.572378641 +0000 UTC m=+129.397114222" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.617248 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" podStartSLOduration=63.617228302 podStartE2EDuration="1m3.617228302s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.593455235 +0000 UTC m=+129.418190816" watchObservedRunningTime="2026-03-09 09:22:08.617228302 +0000 UTC m=+129.441963893" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.618540 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ktw79" podStartSLOduration=63.618529892 podStartE2EDuration="1m3.618529892s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.614142298 +0000 UTC m=+129.438877879" watchObservedRunningTime="2026-03-09 09:22:08.618529892 +0000 UTC m=+129.443265483" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.649213 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.649570 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.149554941 +0000 UTC m=+129.974290522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.666024 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qpk8p" podStartSLOduration=64.666005124 podStartE2EDuration="1m4.666005124s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.642082962 +0000 UTC m=+129.466818553" watchObservedRunningTime="2026-03-09 09:22:08.666005124 +0000 UTC m=+129.490740705" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.686724 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9w27p" podStartSLOduration=63.686696457 podStartE2EDuration="1m3.686696457s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.663327062 +0000 UTC m=+129.488062643" watchObservedRunningTime="2026-03-09 09:22:08.686696457 +0000 UTC m=+129.511432038" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.688606 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" podStartSLOduration=63.688594385 podStartE2EDuration="1m3.688594385s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.685006055 +0000 UTC m=+129.509741636" watchObservedRunningTime="2026-03-09 09:22:08.688594385 +0000 UTC m=+129.513329976" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.708531 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-xsdcf" podStartSLOduration=63.708513904 podStartE2EDuration="1m3.708513904s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:08.70807541 +0000 UTC m=+129.532811001" watchObservedRunningTime="2026-03-09 09:22:08.708513904 +0000 UTC m=+129.533249485" Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.752736 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.752972 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.252938072 +0000 UTC m=+130.077673703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.753153 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.753656 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.253645794 +0000 UTC m=+130.078381385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.854305 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.854699 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.354682664 +0000 UTC m=+130.179418245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:08 crc kubenswrapper[4692]: I0309 09:22:08.955934 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:08 crc kubenswrapper[4692]: E0309 09:22:08.956403 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.456381895 +0000 UTC m=+130.281117676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.057011 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.057251 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.557223339 +0000 UTC m=+130.381958920 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.057517 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.057895 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.557886269 +0000 UTC m=+130.382621850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.128543 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:09 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:09 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:09 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.128594 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.157990 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.158467 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.658451745 +0000 UTC m=+130.483187326 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.260150 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.260548 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.760528527 +0000 UTC m=+130.585264308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.361552 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.361964 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.861949689 +0000 UTC m=+130.686685270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.363773 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" event={"ID":"d59db20e-6720-4cc9-bfd8-b3477c7298cf","Type":"ContainerStarted","Data":"7af20238ea0e31aa07b9529b0189f24882112759a9895e017ced21b3c5bf0842"} Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.367417 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" event={"ID":"03b56c7f-0613-419a-8a32-fe678a4a3e93","Type":"ContainerStarted","Data":"34c8aab210bd8a77dcc7f969ed9754ea1efb365b8ddfd75e8506ef240546cb7d"} Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.370068 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" event={"ID":"93ccefab-c47d-4725-a9ff-55cc50f678f0","Type":"ContainerStarted","Data":"c81530803d2197e9e4ef74165980fa0c90b525f75bbddaac4c866d8d0aff3a8c"} Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.374904 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" event={"ID":"c279ea73-49f4-441d-9a36-bedd8388437e","Type":"ContainerStarted","Data":"ad9d4652573ac6a2ae9b72d310b7ca01bcc1fa2986bed9b3c4f08c5c9555505f"} Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.378746 4692 generic.go:334] "Generic (PLEG): container finished" podID="37e678a0-8c3a-4ff9-9203-9543ddd11bbb" containerID="4af810a4f8dc46992a8ebd2da10e16c78743ffcdc5a059bbd1921d074fc86eef" exitCode=0 Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.378809 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" event={"ID":"37e678a0-8c3a-4ff9-9203-9543ddd11bbb","Type":"ContainerDied","Data":"4af810a4f8dc46992a8ebd2da10e16c78743ffcdc5a059bbd1921d074fc86eef"} Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.390577 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-ftqvs" event={"ID":"f73e489b-1a01-4608-a493-e0ce7dd2c2c5","Type":"ContainerStarted","Data":"fa9404cd67e9772097fc788c10ba65fc14432a14a03687b644573d7958bfcff9"} Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.391982 4692 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rdvhn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392019 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392086 4692 patch_prober.go:28] interesting pod/console-operator-58897d9998-zd4xm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392152 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" podUID="7b9088be-ec8e-43ad-9b73-ae106d9afdc0" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392548 4692 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-kn54j container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392578 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" podUID="700952b9-c6cb-403f-855b-b3cbd855e845" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392630 4692 patch_prober.go:28] interesting pod/downloads-7954f5f757-snq8b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392651 4692 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-29jhq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392659 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-snq8b" podUID="aa908134-de3c-43e0-ace1-94fc548861c8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.392675 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" podUID="9c16e62e-9744-45c8-88d9-0c289df743a1" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.411407 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" podStartSLOduration=65.411391471 podStartE2EDuration="1m5.411391471s" podCreationTimestamp="2026-03-09 09:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:09.411067501 +0000 UTC m=+130.235803072" watchObservedRunningTime="2026-03-09 09:22:09.411391471 +0000 UTC m=+130.236127052" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.413544 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rfkdj" podStartSLOduration=64.413536956 podStartE2EDuration="1m4.413536956s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:09.390970126 +0000 UTC m=+130.215705727" watchObservedRunningTime="2026-03-09 09:22:09.413536956 +0000 UTC m=+130.238272537" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.463266 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.465755 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:09.965741853 +0000 UTC m=+130.790477434 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.475868 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-vdjpl" podStartSLOduration=64.475852422 podStartE2EDuration="1m4.475852422s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:09.474763169 +0000 UTC m=+130.299498770" watchObservedRunningTime="2026-03-09 09:22:09.475852422 +0000 UTC m=+130.300587993" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.568501 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.570728 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.070688543 +0000 UTC m=+130.895424124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.673385 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.673709 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.173695473 +0000 UTC m=+130.998431054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.774425 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.774735 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.274718063 +0000 UTC m=+131.099453644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.876095 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.876560 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.376544217 +0000 UTC m=+131.201279798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.963621 4692 ???:1] "http: TLS handshake error from 192.168.126.11:51062: no serving certificate available for the kubelet" Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.977633 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.977800 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.477776343 +0000 UTC m=+131.302511924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:09 crc kubenswrapper[4692]: I0309 09:22:09.977914 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:09 crc kubenswrapper[4692]: E0309 09:22:09.978207 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.478194566 +0000 UTC m=+131.302930147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.010235 4692 ???:1] "http: TLS handshake error from 192.168.126.11:51068: no serving certificate available for the kubelet" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.050456 4692 ???:1] "http: TLS handshake error from 192.168.126.11:51078: no serving certificate available for the kubelet" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.078506 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.078784 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.578767822 +0000 UTC m=+131.403503403 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.101517 4692 ???:1] "http: TLS handshake error from 192.168.126.11:51086: no serving certificate available for the kubelet" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.127439 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:10 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:10 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:10 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.127499 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.180226 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.180713 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.6806976 +0000 UTC m=+131.505433181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.194536 4692 ???:1] "http: TLS handshake error from 192.168.126.11:51096: no serving certificate available for the kubelet" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.224625 4692 ???:1] "http: TLS handshake error from 192.168.126.11:51102: no serving certificate available for the kubelet" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.281388 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.281724 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.781708879 +0000 UTC m=+131.606444460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.318240 4692 ???:1] "http: TLS handshake error from 192.168.126.11:51114: no serving certificate available for the kubelet" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.382367 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.382747 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.882729069 +0000 UTC m=+131.707464850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.401995 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" event={"ID":"37e678a0-8c3a-4ff9-9203-9543ddd11bbb","Type":"ContainerStarted","Data":"f6d70def238f57b30eec34d7d055ade43f937e9d514ee05fade3a0ee951f4e61"} Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.434460 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" podStartSLOduration=65.43444477 podStartE2EDuration="1m5.43444477s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:10.432278884 +0000 UTC m=+131.257014475" watchObservedRunningTime="2026-03-09 09:22:10.43444477 +0000 UTC m=+131.259180351" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.435338 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-29jhq" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.484373 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.485359 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:10.985343437 +0000 UTC m=+131.810079018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.585957 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.586426 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.086407288 +0000 UTC m=+131.911142869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.665350 4692 ???:1] "http: TLS handshake error from 192.168.126.11:53444: no serving certificate available for the kubelet" Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.687321 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.687498 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.187473079 +0000 UTC m=+132.012208660 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.687714 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.688009 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.188001395 +0000 UTC m=+132.012736976 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.796796 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.796999 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.296971588 +0000 UTC m=+132.121707169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.797046 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.797422 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.297409631 +0000 UTC m=+132.122145422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.898136 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.898309 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.398283226 +0000 UTC m=+132.223018807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.898642 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.898970 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.398962707 +0000 UTC m=+132.223698288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.999539 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:10 crc kubenswrapper[4692]: E0309 09:22:10.999794 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.49976911 +0000 UTC m=+132.324504691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:10 crc kubenswrapper[4692]: I0309 09:22:10.999954 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.000370 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.500355228 +0000 UTC m=+132.325090999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.100803 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.100996 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.600970735 +0000 UTC m=+132.425706316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.101123 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.101416 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.601408489 +0000 UTC m=+132.426144060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.202440 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.202618 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.702593192 +0000 UTC m=+132.527328773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.202696 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.203028 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.703015565 +0000 UTC m=+132.527751146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.231663 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:11 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:11 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:11 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.231723 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.290987 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-kn54j" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.304607 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.305015 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.805001754 +0000 UTC m=+132.629737335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.354567 4692 ???:1] "http: TLS handshake error from 192.168.126.11:53460: no serving certificate available for the kubelet" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.405778 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.406098 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:11.906087886 +0000 UTC m=+132.730823467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.416151 4692 generic.go:334] "Generic (PLEG): container finished" podID="f96b54e5-804b-4d9e-9df9-538352c52073" containerID="ac247f9bfc4074d231deb121d854bad9b6a80025df752a0138fad95d1a0dbfb0" exitCode=0 Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.416284 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" event={"ID":"f96b54e5-804b-4d9e-9df9-538352c52073","Type":"ContainerDied","Data":"ac247f9bfc4074d231deb121d854bad9b6a80025df752a0138fad95d1a0dbfb0"} Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.422014 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" event={"ID":"03b56c7f-0613-419a-8a32-fe678a4a3e93","Type":"ContainerStarted","Data":"231fd280374f6023b3688a61b9c4d4643a5986abe029492bca80b7010a080ce5"} Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.422347 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.450991 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4bqhb"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.457446 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4bqhb"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.457572 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.468857 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.506607 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.506840 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6gqq\" (UniqueName: \"kubernetes.io/projected/73819651-942b-4436-b0b1-7225cfb20ef1-kube-api-access-g6gqq\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.506887 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-catalog-content\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.506935 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-utilities\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.507207 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.007184638 +0000 UTC m=+132.831920369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.607630 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6gqq\" (UniqueName: \"kubernetes.io/projected/73819651-942b-4436-b0b1-7225cfb20ef1-kube-api-access-g6gqq\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.607955 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-catalog-content\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.607992 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-utilities\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.608051 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.608330 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.108319491 +0000 UTC m=+132.933055062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.608538 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-catalog-content\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.608572 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-utilities\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.636449 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6gqq\" (UniqueName: \"kubernetes.io/projected/73819651-942b-4436-b0b1-7225cfb20ef1-kube-api-access-g6gqq\") pod \"certified-operators-4bqhb\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.643919 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ktmgp"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.645313 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.648098 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.664288 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ktmgp"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.697224 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nxm8t"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.697476 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" podUID="1e53fbe9-5027-4dae-8f6f-f2eda177f94b" containerName="controller-manager" containerID="cri-o://02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26" gracePeriod=30 Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.711259 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.711418 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-utilities\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.711444 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gkcn\" (UniqueName: \"kubernetes.io/projected/067264d3-8fc2-46cd-8d94-e89f9e25a500-kube-api-access-6gkcn\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.711465 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-catalog-content\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.711646 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.21162208 +0000 UTC m=+133.036357661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.714716 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.714970 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" podUID="4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" containerName="route-controller-manager" containerID="cri-o://c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11" gracePeriod=30 Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.778396 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.813868 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.814238 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-utilities\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.814266 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gkcn\" (UniqueName: \"kubernetes.io/projected/067264d3-8fc2-46cd-8d94-e89f9e25a500-kube-api-access-6gkcn\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.814287 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-catalog-content\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.814735 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-catalog-content\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.814858 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-utilities\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.815216 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.315198368 +0000 UTC m=+133.139934149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.849813 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gkcn\" (UniqueName: \"kubernetes.io/projected/067264d3-8fc2-46cd-8d94-e89f9e25a500-kube-api-access-6gkcn\") pod \"community-operators-ktmgp\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.871235 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qgmsd"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.872614 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.872870 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qgmsd"] Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.922563 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.922796 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-utilities\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.922848 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgtzd\" (UniqueName: \"kubernetes.io/projected/2f27156a-57b6-4230-aa9b-1809cbfb3569-kube-api-access-jgtzd\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.922902 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-catalog-content\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:11 crc kubenswrapper[4692]: E0309 09:22:11.923110 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.423095758 +0000 UTC m=+133.247831339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.980509 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:22:11 crc kubenswrapper[4692]: I0309 09:22:11.999416 4692 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.024146 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.024211 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-utilities\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.024238 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgtzd\" (UniqueName: \"kubernetes.io/projected/2f27156a-57b6-4230-aa9b-1809cbfb3569-kube-api-access-jgtzd\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.024289 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-catalog-content\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.024693 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-catalog-content\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.024916 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.524905152 +0000 UTC m=+133.349640733 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.025237 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-utilities\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.042068 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q9nxd"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.045779 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.052951 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgtzd\" (UniqueName: \"kubernetes.io/projected/2f27156a-57b6-4230-aa9b-1809cbfb3569-kube-api-access-jgtzd\") pod \"certified-operators-qgmsd\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.058464 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q9nxd"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.125701 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.126008 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-catalog-content\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.126076 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9w5h\" (UniqueName: \"kubernetes.io/projected/27500a21-baed-4a7c-9af8-e616fee71857-kube-api-access-w9w5h\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.126107 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-utilities\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.126239 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.626223461 +0000 UTC m=+133.450959042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.136390 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:12 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:12 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:12 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.136451 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.219909 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.222940 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.231546 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.231608 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-catalog-content\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.231681 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9w5h\" (UniqueName: \"kubernetes.io/projected/27500a21-baed-4a7c-9af8-e616fee71857-kube-api-access-w9w5h\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.231705 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-utilities\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.232120 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-utilities\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.232358 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-catalog-content\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.232587 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-09 09:22:12.732573133 +0000 UTC m=+133.557308924 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxwb2" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.257129 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9w5h\" (UniqueName: \"kubernetes.io/projected/27500a21-baed-4a7c-9af8-e616fee71857-kube-api-access-w9w5h\") pod \"community-operators-q9nxd\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.264260 4692 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-09T09:22:11.999484805Z","Handler":null,"Name":""} Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.284288 4692 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.284317 4692 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.299948 4692 ???:1] "http: TLS handshake error from 192.168.126.11:53470: no serving certificate available for the kubelet" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.308720 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.334393 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-client-ca\") pod \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.334454 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-serving-cert\") pod \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.334495 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mwq5\" (UniqueName: \"kubernetes.io/projected/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-kube-api-access-4mwq5\") pod \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.334608 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-config\") pod \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\" (UID: \"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.334753 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.335195 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-client-ca" (OuterVolumeSpecName: "client-ca") pod "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" (UID: "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.335594 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-config" (OuterVolumeSpecName: "config") pod "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" (UID: "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.342929 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" (UID: "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.343265 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.343409 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-kube-api-access-4mwq5" (OuterVolumeSpecName: "kube-api-access-4mwq5") pod "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" (UID: "4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb"). InnerVolumeSpecName "kube-api-access-4mwq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.378984 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.432100 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" event={"ID":"03b56c7f-0613-419a-8a32-fe678a4a3e93","Type":"ContainerStarted","Data":"d160c84c0e35d93a0a67ab40123a6aa8fbd91390201f2fe2834bfe74f8074b07"} Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.432140 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" event={"ID":"03b56c7f-0613-419a-8a32-fe678a4a3e93","Type":"ContainerStarted","Data":"d1d159d968fcd751658137255cf84e7a9e2e4b7412e8ecd645329c0d40e89f2f"} Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.438093 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-proxy-ca-bundles\") pod \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.438152 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-config\") pod \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.438283 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-client-ca\") pod \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.438350 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22vg9\" (UniqueName: \"kubernetes.io/projected/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-kube-api-access-22vg9\") pod \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.439042 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-serving-cert\") pod \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\" (UID: \"1e53fbe9-5027-4dae-8f6f-f2eda177f94b\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.439279 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-client-ca" (OuterVolumeSpecName: "client-ca") pod "1e53fbe9-5027-4dae-8f6f-f2eda177f94b" (UID: "1e53fbe9-5027-4dae-8f6f-f2eda177f94b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.439293 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.441970 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.442001 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.442014 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.442031 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.442044 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mwq5\" (UniqueName: \"kubernetes.io/projected/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb-kube-api-access-4mwq5\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.442687 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-config" (OuterVolumeSpecName: "config") pod "1e53fbe9-5027-4dae-8f6f-f2eda177f94b" (UID: "1e53fbe9-5027-4dae-8f6f-f2eda177f94b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.443337 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1e53fbe9-5027-4dae-8f6f-f2eda177f94b" (UID: "1e53fbe9-5027-4dae-8f6f-f2eda177f94b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.445043 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-kube-api-access-22vg9" (OuterVolumeSpecName: "kube-api-access-22vg9") pod "1e53fbe9-5027-4dae-8f6f-f2eda177f94b" (UID: "1e53fbe9-5027-4dae-8f6f-f2eda177f94b"). InnerVolumeSpecName "kube-api-access-22vg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.447069 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1e53fbe9-5027-4dae-8f6f-f2eda177f94b" (UID: "1e53fbe9-5027-4dae-8f6f-f2eda177f94b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.452435 4692 generic.go:334] "Generic (PLEG): container finished" podID="4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" containerID="c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11" exitCode=0 Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.452874 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.454466 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" event={"ID":"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb","Type":"ContainerDied","Data":"c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11"} Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.454519 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z" event={"ID":"4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb","Type":"ContainerDied","Data":"3a9d0ad7db8c22a125ddd7d08097b231dc348773f75fb465e29000d2ae09f95b"} Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.454535 4692 scope.go:117] "RemoveContainer" containerID="c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.462732 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-jdvnz" podStartSLOduration=14.462708532 podStartE2EDuration="14.462708532s" podCreationTimestamp="2026-03-09 09:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:12.455027407 +0000 UTC m=+133.279762988" watchObservedRunningTime="2026-03-09 09:22:12.462708532 +0000 UTC m=+133.287444113" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.473039 4692 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.473083 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.482542 4692 generic.go:334] "Generic (PLEG): container finished" podID="1e53fbe9-5027-4dae-8f6f-f2eda177f94b" containerID="02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26" exitCode=0 Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.482631 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.482671 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" event={"ID":"1e53fbe9-5027-4dae-8f6f-f2eda177f94b","Type":"ContainerDied","Data":"02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26"} Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.482963 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nxm8t" event={"ID":"1e53fbe9-5027-4dae-8f6f-f2eda177f94b","Type":"ContainerDied","Data":"04e00af1f96d5952625bde42114285e09711fbc65d39383ef21157e084f1acf7"} Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.495088 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bkn6" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.501297 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.509450 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-q286z"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.509395 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxwb2\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.527991 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4bqhb"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.542563 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nxm8t"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.543500 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22vg9\" (UniqueName: \"kubernetes.io/projected/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-kube-api-access-22vg9\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.543513 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.543524 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.543552 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e53fbe9-5027-4dae-8f6f-f2eda177f94b-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.545343 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nxm8t"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.553045 4692 scope.go:117] "RemoveContainer" containerID="c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11" Mar 09 09:22:12 crc kubenswrapper[4692]: W0309 09:22:12.554128 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73819651_942b_4436_b0b1_7225cfb20ef1.slice/crio-5571d809348713925825502d029a7db47086da2659bfee1591e54bf18709d5c1 WatchSource:0}: Error finding container 5571d809348713925825502d029a7db47086da2659bfee1591e54bf18709d5c1: Status 404 returned error can't find the container with id 5571d809348713925825502d029a7db47086da2659bfee1591e54bf18709d5c1 Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.554239 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11\": container with ID starting with c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11 not found: ID does not exist" containerID="c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.554281 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11"} err="failed to get container status \"c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11\": rpc error: code = NotFound desc = could not find container \"c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11\": container with ID starting with c394117465a7bf02f2a55fee318d8a6f93f2393213298ac28928a2063a5e9d11 not found: ID does not exist" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.554308 4692 scope.go:117] "RemoveContainer" containerID="02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.584339 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qgmsd"] Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.585267 4692 scope.go:117] "RemoveContainer" containerID="02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26" Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.586208 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26\": container with ID starting with 02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26 not found: ID does not exist" containerID="02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.586246 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26"} err="failed to get container status \"02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26\": rpc error: code = NotFound desc = could not find container \"02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26\": container with ID starting with 02c2bb45feafee52327aaa71f5d86b859da1b0b1e0628eb6948ac6be1baa8f26 not found: ID does not exist" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.587087 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.654498 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ktmgp"] Mar 09 09:22:12 crc kubenswrapper[4692]: W0309 09:22:12.670065 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod067264d3_8fc2_46cd_8d94_e89f9e25a500.slice/crio-7b70955e661f01296c52dec402d93d90f896b537a065ed4a62baf31ed4e6fc31 WatchSource:0}: Error finding container 7b70955e661f01296c52dec402d93d90f896b537a065ed4a62baf31ed4e6fc31: Status 404 returned error can't find the container with id 7b70955e661f01296c52dec402d93d90f896b537a065ed4a62baf31ed4e6fc31 Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.674356 4692 ???:1] "http: TLS handshake error from 192.168.126.11:53474: no serving certificate available for the kubelet" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.785593 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.797713 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.806542 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.848554 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume\") pod \"f96b54e5-804b-4d9e-9df9-538352c52073\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.848611 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96b54e5-804b-4d9e-9df9-538352c52073-config-volume\") pod \"f96b54e5-804b-4d9e-9df9-538352c52073\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.848651 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfk64\" (UniqueName: \"kubernetes.io/projected/f96b54e5-804b-4d9e-9df9-538352c52073-kube-api-access-cfk64\") pod \"f96b54e5-804b-4d9e-9df9-538352c52073\" (UID: \"f96b54e5-804b-4d9e-9df9-538352c52073\") " Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.849601 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f96b54e5-804b-4d9e-9df9-538352c52073-config-volume" (OuterVolumeSpecName: "config-volume") pod "f96b54e5-804b-4d9e-9df9-538352c52073" (UID: "f96b54e5-804b-4d9e-9df9-538352c52073"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.853128 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f96b54e5-804b-4d9e-9df9-538352c52073" (UID: "f96b54e5-804b-4d9e-9df9-538352c52073"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.853788 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f96b54e5-804b-4d9e-9df9-538352c52073-kube-api-access-cfk64" (OuterVolumeSpecName: "kube-api-access-cfk64") pod "f96b54e5-804b-4d9e-9df9-538352c52073" (UID: "f96b54e5-804b-4d9e-9df9-538352c52073"). InnerVolumeSpecName "kube-api-access-cfk64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.853820 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.855250 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.856557 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:12 crc kubenswrapper[4692]: E0309 09:22:12.856633 4692 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.950358 4692 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f96b54e5-804b-4d9e-9df9-538352c52073-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.950389 4692 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f96b54e5-804b-4d9e-9df9-538352c52073-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.950401 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfk64\" (UniqueName: \"kubernetes.io/projected/f96b54e5-804b-4d9e-9df9-538352c52073-kube-api-access-cfk64\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:12 crc kubenswrapper[4692]: I0309 09:22:12.996787 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q9nxd"] Mar 09 09:22:13 crc kubenswrapper[4692]: W0309 09:22:13.020658 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27500a21_baed_4a7c_9af8_e616fee71857.slice/crio-c2dd6eb569055c8b33157bfdf8bb6206a92145ba93cf980592e31347c6c2f70a WatchSource:0}: Error finding container c2dd6eb569055c8b33157bfdf8bb6206a92145ba93cf980592e31347c6c2f70a: Status 404 returned error can't find the container with id c2dd6eb569055c8b33157bfdf8bb6206a92145ba93cf980592e31347c6c2f70a Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.043169 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxwb2"] Mar 09 09:22:13 crc kubenswrapper[4692]: W0309 09:22:13.061629 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod946e899e_f89e_4f66_80a2_1f3effc06b7f.slice/crio-2302f75aacff21fe204ca67e0ed8bc3dd242d7e995da02f4181eec23548af690 WatchSource:0}: Error finding container 2302f75aacff21fe204ca67e0ed8bc3dd242d7e995da02f4181eec23548af690: Status 404 returned error can't find the container with id 2302f75aacff21fe204ca67e0ed8bc3dd242d7e995da02f4181eec23548af690 Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.125310 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.130895 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:13 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:13 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:13 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.130971 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.290329 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-549d7bf89b-9sxvg"] Mar 09 09:22:13 crc kubenswrapper[4692]: E0309 09:22:13.290941 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e53fbe9-5027-4dae-8f6f-f2eda177f94b" containerName="controller-manager" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.290960 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e53fbe9-5027-4dae-8f6f-f2eda177f94b" containerName="controller-manager" Mar 09 09:22:13 crc kubenswrapper[4692]: E0309 09:22:13.290972 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" containerName="route-controller-manager" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.290978 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" containerName="route-controller-manager" Mar 09 09:22:13 crc kubenswrapper[4692]: E0309 09:22:13.290985 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f96b54e5-804b-4d9e-9df9-538352c52073" containerName="collect-profiles" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.290992 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f96b54e5-804b-4d9e-9df9-538352c52073" containerName="collect-profiles" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.291112 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f96b54e5-804b-4d9e-9df9-538352c52073" containerName="collect-profiles" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.291127 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" containerName="route-controller-manager" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.291139 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e53fbe9-5027-4dae-8f6f-f2eda177f94b" containerName="controller-manager" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.291532 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.294989 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.296117 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.299460 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.303238 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.299548 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.299599 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.305307 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.308868 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.309399 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.309800 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.310203 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.310487 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.312058 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.322540 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.330692 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-549d7bf89b-9sxvg"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.342940 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.343435 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355439 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-config\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355479 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-proxy-ca-bundles\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355505 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-config\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355522 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-752bq\" (UniqueName: \"kubernetes.io/projected/934d185d-4068-409d-8e31-6b8f63ca6e48-kube-api-access-752bq\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355558 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9b0841d-ef1d-4057-b4eb-b136af058aab-serving-cert\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355587 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2ldl\" (UniqueName: \"kubernetes.io/projected/f9b0841d-ef1d-4057-b4eb-b136af058aab-kube-api-access-j2ldl\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355608 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-client-ca\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355622 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-client-ca\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.355652 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/934d185d-4068-409d-8e31-6b8f63ca6e48-serving-cert\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.422074 4692 patch_prober.go:28] interesting pod/downloads-7954f5f757-snq8b container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.422365 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-snq8b" podUID="aa908134-de3c-43e0-ace1-94fc548861c8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.422135 4692 patch_prober.go:28] interesting pod/downloads-7954f5f757-snq8b container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.423185 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-snq8b" podUID="aa908134-de3c-43e0-ace1-94fc548861c8" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.440967 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.441022 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.442306 4692 patch_prober.go:28] interesting pod/console-f9d7485db-rddqb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.442455 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rddqb" podUID="7b8b20b2-f067-4abd-adc3-7b1a2bf36742" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456394 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2ldl\" (UniqueName: \"kubernetes.io/projected/f9b0841d-ef1d-4057-b4eb-b136af058aab-kube-api-access-j2ldl\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456470 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-client-ca\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456497 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-client-ca\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456544 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/934d185d-4068-409d-8e31-6b8f63ca6e48-serving-cert\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456585 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-config\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456601 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-proxy-ca-bundles\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456621 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-config\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456638 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-752bq\" (UniqueName: \"kubernetes.io/projected/934d185d-4068-409d-8e31-6b8f63ca6e48-kube-api-access-752bq\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.456673 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9b0841d-ef1d-4057-b4eb-b136af058aab-serving-cert\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.460049 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-client-ca\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.462495 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-config\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.463796 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-config\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.463895 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-client-ca\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.472200 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9b0841d-ef1d-4057-b4eb-b136af058aab-serving-cert\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.473005 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-proxy-ca-bundles\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.473708 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/934d185d-4068-409d-8e31-6b8f63ca6e48-serving-cert\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.491344 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2ldl\" (UniqueName: \"kubernetes.io/projected/f9b0841d-ef1d-4057-b4eb-b136af058aab-kube-api-access-j2ldl\") pod \"controller-manager-549d7bf89b-9sxvg\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.493572 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-752bq\" (UniqueName: \"kubernetes.io/projected/934d185d-4068-409d-8e31-6b8f63ca6e48-kube-api-access-752bq\") pod \"route-controller-manager-8f6f7fb64-8wvwf\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.494050 4692 generic.go:334] "Generic (PLEG): container finished" podID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerID="3ed3288e4cde1e8118367b9fac99d39dcb77cb82dc3e097c1ce2566f60a7a2bd" exitCode=0 Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.494219 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktmgp" event={"ID":"067264d3-8fc2-46cd-8d94-e89f9e25a500","Type":"ContainerDied","Data":"3ed3288e4cde1e8118367b9fac99d39dcb77cb82dc3e097c1ce2566f60a7a2bd"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.494330 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktmgp" event={"ID":"067264d3-8fc2-46cd-8d94-e89f9e25a500","Type":"ContainerStarted","Data":"7b70955e661f01296c52dec402d93d90f896b537a065ed4a62baf31ed4e6fc31"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.496512 4692 generic.go:334] "Generic (PLEG): container finished" podID="27500a21-baed-4a7c-9af8-e616fee71857" containerID="75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c" exitCode=0 Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.496552 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9nxd" event={"ID":"27500a21-baed-4a7c-9af8-e616fee71857","Type":"ContainerDied","Data":"75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.496603 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9nxd" event={"ID":"27500a21-baed-4a7c-9af8-e616fee71857","Type":"ContainerStarted","Data":"c2dd6eb569055c8b33157bfdf8bb6206a92145ba93cf980592e31347c6c2f70a"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.507575 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.508484 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv" event={"ID":"f96b54e5-804b-4d9e-9df9-538352c52073","Type":"ContainerDied","Data":"5307770c0427ac67f58c5e90e32674d8e3abf5cbe12c7c389ef27085cfa41d13"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.508530 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5307770c0427ac67f58c5e90e32674d8e3abf5cbe12c7c389ef27085cfa41d13" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.515619 4692 generic.go:334] "Generic (PLEG): container finished" podID="73819651-942b-4436-b0b1-7225cfb20ef1" containerID="57f97e200a45c7ffe5ff133937b61b7a6a5d987f6f2c7bdf4d137d62ad873142" exitCode=0 Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.515794 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4bqhb" event={"ID":"73819651-942b-4436-b0b1-7225cfb20ef1","Type":"ContainerDied","Data":"57f97e200a45c7ffe5ff133937b61b7a6a5d987f6f2c7bdf4d137d62ad873142"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.515834 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4bqhb" event={"ID":"73819651-942b-4436-b0b1-7225cfb20ef1","Type":"ContainerStarted","Data":"5571d809348713925825502d029a7db47086da2659bfee1591e54bf18709d5c1"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.519846 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" event={"ID":"946e899e-f89e-4f66-80a2-1f3effc06b7f","Type":"ContainerStarted","Data":"9d8f07d1dd0caa34ab1801883390d68844fd2f19b56dee0acca723bd3cd9823d"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.519897 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" event={"ID":"946e899e-f89e-4f66-80a2-1f3effc06b7f","Type":"ContainerStarted","Data":"2302f75aacff21fe204ca67e0ed8bc3dd242d7e995da02f4181eec23548af690"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.520652 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.528996 4692 generic.go:334] "Generic (PLEG): container finished" podID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerID="411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d" exitCode=0 Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.529281 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qgmsd" event={"ID":"2f27156a-57b6-4230-aa9b-1809cbfb3569","Type":"ContainerDied","Data":"411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.529339 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qgmsd" event={"ID":"2f27156a-57b6-4230-aa9b-1809cbfb3569","Type":"ContainerStarted","Data":"2fd137a190bddd010c8c6e2c625b368d2503c19f62836ace20067310dd96db5b"} Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.567942 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.568567 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.571395 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.576994 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.582504 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" podStartSLOduration=68.582486479 podStartE2EDuration="1m8.582486479s" podCreationTimestamp="2026-03-09 09:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:13.579955482 +0000 UTC m=+134.404691063" watchObservedRunningTime="2026-03-09 09:22:13.582486479 +0000 UTC m=+134.407222060" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.582929 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.621538 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.632542 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.655785 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vkkhr"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.656976 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.662208 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.662443 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.662551 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.673575 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkkhr"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.717285 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.717975 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.723180 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.727562 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.727763 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.763778 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-catalog-content\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.763830 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzw78\" (UniqueName: \"kubernetes.io/projected/3d4b7447-a6e6-4530-b529-2ec461452791-kube-api-access-nzw78\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.763857 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-utilities\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.763922 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.763960 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.763982 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.764003 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.764392 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.785175 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.801064 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-26mfz" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.809068 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.865140 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.865521 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.865579 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-catalog-content\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.865611 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzw78\" (UniqueName: \"kubernetes.io/projected/3d4b7447-a6e6-4530-b529-2ec461452791-kube-api-access-nzw78\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.865632 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-utilities\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.866058 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-utilities\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.866100 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.866707 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-catalog-content\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.889214 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzw78\" (UniqueName: \"kubernetes.io/projected/3d4b7447-a6e6-4530-b529-2ec461452791-kube-api-access-nzw78\") pod \"redhat-marketplace-vkkhr\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.890895 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.893608 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.893673 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-549d7bf89b-9sxvg"] Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.921803 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.983702 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:22:13 crc kubenswrapper[4692]: I0309 09:22:13.990374 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf"] Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.037689 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8vbb6"] Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.038670 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.047876 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vbb6"] Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.065864 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.068085 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-catalog-content\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.068124 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-utilities\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.068279 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhsds\" (UniqueName: \"kubernetes.io/projected/66179a0b-d877-46fb-a854-5426993f4b36-kube-api-access-qhsds\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.080511 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e53fbe9-5027-4dae-8f6f-f2eda177f94b" path="/var/lib/kubelet/pods/1e53fbe9-5027-4dae-8f6f-f2eda177f94b/volumes" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.081369 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb" path="/var/lib/kubelet/pods/4b20b5e8-11a3-49ef-9b04-4b5a5bc48adb/volumes" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.082323 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.126619 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:14 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:14 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:14 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.126679 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.169188 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-catalog-content\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.169238 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-utilities\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.169303 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhsds\" (UniqueName: \"kubernetes.io/projected/66179a0b-d877-46fb-a854-5426993f4b36-kube-api-access-qhsds\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.170274 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-catalog-content\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.170526 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-utilities\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.187473 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhsds\" (UniqueName: \"kubernetes.io/projected/66179a0b-d877-46fb-a854-5426993f4b36-kube-api-access-qhsds\") pod \"redhat-marketplace-8vbb6\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.241563 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.241627 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.248396 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.363034 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.535479 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vbp6d" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.548662 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-4sdrj" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.613429 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-zd4xm" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.659294 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pnf7m"] Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.660614 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.667338 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.695390 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pnf7m"] Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.788680 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-utilities\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.788792 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-catalog-content\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.788881 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pnq6\" (UniqueName: \"kubernetes.io/projected/15d06ca2-19f5-4c88-8c32-51e491ffa834-kube-api-access-2pnq6\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.889715 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-utilities\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.889775 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-catalog-content\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.890000 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pnq6\" (UniqueName: \"kubernetes.io/projected/15d06ca2-19f5-4c88-8c32-51e491ffa834-kube-api-access-2pnq6\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.892404 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-catalog-content\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.893245 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-utilities\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.908099 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pnq6\" (UniqueName: \"kubernetes.io/projected/15d06ca2-19f5-4c88-8c32-51e491ffa834-kube-api-access-2pnq6\") pod \"redhat-operators-pnf7m\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:14 crc kubenswrapper[4692]: I0309 09:22:14.984721 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.038329 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-998gk"] Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.039323 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.049905 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-998gk"] Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.127518 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:15 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:15 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:15 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.127585 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.199963 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-catalog-content\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.200129 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t5jn\" (UniqueName: \"kubernetes.io/projected/fd840400-6e1f-4556-b60b-0cbacff31f1d-kube-api-access-4t5jn\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.200182 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-utilities\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.301990 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-catalog-content\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.302397 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-catalog-content\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.302463 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-utilities\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.302480 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t5jn\" (UniqueName: \"kubernetes.io/projected/fd840400-6e1f-4556-b60b-0cbacff31f1d-kube-api-access-4t5jn\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.302725 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-utilities\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.340067 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t5jn\" (UniqueName: \"kubernetes.io/projected/fd840400-6e1f-4556-b60b-0cbacff31f1d-kube-api-access-4t5jn\") pod \"redhat-operators-998gk\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:15 crc kubenswrapper[4692]: I0309 09:22:15.353931 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:22:16 crc kubenswrapper[4692]: I0309 09:22:16.126794 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:16 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:16 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:16 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:16 crc kubenswrapper[4692]: I0309 09:22:16.127141 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:17 crc kubenswrapper[4692]: I0309 09:22:17.126365 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:17 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:17 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:17 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:17 crc kubenswrapper[4692]: I0309 09:22:17.126447 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:17 crc kubenswrapper[4692]: I0309 09:22:17.667840 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:22:17 crc kubenswrapper[4692]: I0309 09:22:17.824923 4692 ???:1] "http: TLS handshake error from 192.168.126.11:53478: no serving certificate available for the kubelet" Mar 09 09:22:18 crc kubenswrapper[4692]: I0309 09:22:18.086677 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 09 09:22:18 crc kubenswrapper[4692]: I0309 09:22:18.126936 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:18 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:18 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:18 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:18 crc kubenswrapper[4692]: I0309 09:22:18.127009 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:19 crc kubenswrapper[4692]: I0309 09:22:19.127020 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:19 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:19 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:19 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:19 crc kubenswrapper[4692]: I0309 09:22:19.127752 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:19 crc kubenswrapper[4692]: W0309 09:22:19.194849 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod934d185d_4068_409d_8e31_6b8f63ca6e48.slice/crio-56cd50398bb4dee57f9f5c41d9258c219a85a789e4aca7fb16193cfc132a0dbb WatchSource:0}: Error finding container 56cd50398bb4dee57f9f5c41d9258c219a85a789e4aca7fb16193cfc132a0dbb: Status 404 returned error can't find the container with id 56cd50398bb4dee57f9f5c41d9258c219a85a789e4aca7fb16193cfc132a0dbb Mar 09 09:22:19 crc kubenswrapper[4692]: I0309 09:22:19.570038 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" event={"ID":"934d185d-4068-409d-8e31-6b8f63ca6e48","Type":"ContainerStarted","Data":"56cd50398bb4dee57f9f5c41d9258c219a85a789e4aca7fb16193cfc132a0dbb"} Mar 09 09:22:19 crc kubenswrapper[4692]: I0309 09:22:19.572524 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" event={"ID":"f9b0841d-ef1d-4057-b4eb-b136af058aab","Type":"ContainerStarted","Data":"a84f1cfbe073fe80c0aa6f75d63b0a2119821d39be5840af9486a37c4e74499c"} Mar 09 09:22:20 crc kubenswrapper[4692]: I0309 09:22:20.125706 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:20 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:20 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:20 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:20 crc kubenswrapper[4692]: I0309 09:22:20.125813 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:20 crc kubenswrapper[4692]: I0309 09:22:20.142236 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=2.142216381 podStartE2EDuration="2.142216381s" podCreationTimestamp="2026-03-09 09:22:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:20.140614482 +0000 UTC m=+140.965350073" watchObservedRunningTime="2026-03-09 09:22:20.142216381 +0000 UTC m=+140.966951962" Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.127333 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:21 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:21 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:21 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.127635 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.590886 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" event={"ID":"934d185d-4068-409d-8e31-6b8f63ca6e48","Type":"ContainerStarted","Data":"81053e36c98e67cd66fb5aed96d9aafda2635780f331e2dc02d1b8c129780904"} Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.592560 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.595298 4692 patch_prober.go:28] interesting pod/route-controller-manager-8f6f7fb64-8wvwf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" start-of-body= Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.595346 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.618593 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" podStartSLOduration=10.618568114 podStartE2EDuration="10.618568114s" podCreationTimestamp="2026-03-09 09:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:21.6135467 +0000 UTC m=+142.438282291" watchObservedRunningTime="2026-03-09 09:22:21.618568114 +0000 UTC m=+142.443303705" Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.780467 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vbb6"] Mar 09 09:22:21 crc kubenswrapper[4692]: I0309 09:22:21.788754 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pnf7m"] Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.107461 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkkhr"] Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.128181 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.131242 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.131591 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:22 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:22 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:22 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.131657 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.202592 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-998gk"] Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.612186 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" event={"ID":"f9b0841d-ef1d-4057-b4eb-b136af058aab","Type":"ContainerStarted","Data":"4c04c043b9a5866506ea255353446b8d73f2189e127dcd71784749b9482bf9fd"} Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.616898 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.631673 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550802-w82g5" event={"ID":"64c0a3f4-29f0-42fa-83bc-83e15171ed77","Type":"ContainerStarted","Data":"b861a5a13ae14d622c134d6b2ea01cde37bbfae9efc6df4d37057cf71b528c86"} Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.633457 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.643547 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.643808 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" podStartSLOduration=11.643791338 podStartE2EDuration="11.643791338s" podCreationTimestamp="2026-03-09 09:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:22.641741486 +0000 UTC m=+143.466477077" watchObservedRunningTime="2026-03-09 09:22:22.643791338 +0000 UTC m=+143.468526919" Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.699608 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29550802-w82g5" podStartSLOduration=4.846325853 podStartE2EDuration="22.699585865s" podCreationTimestamp="2026-03-09 09:22:00 +0000 UTC" firstStartedPulling="2026-03-09 09:22:03.576434085 +0000 UTC m=+124.401169656" lastFinishedPulling="2026-03-09 09:22:21.429694087 +0000 UTC m=+142.254429668" observedRunningTime="2026-03-09 09:22:22.681370488 +0000 UTC m=+143.506106079" watchObservedRunningTime="2026-03-09 09:22:22.699585865 +0000 UTC m=+143.524321446" Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.709561 4692 csr.go:261] certificate signing request csr-cstmx is approved, waiting to be issued Mar 09 09:22:22 crc kubenswrapper[4692]: I0309 09:22:22.716316 4692 csr.go:257] certificate signing request csr-cstmx is issued Mar 09 09:22:22 crc kubenswrapper[4692]: E0309 09:22:22.856399 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:22 crc kubenswrapper[4692]: E0309 09:22:22.858752 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:22 crc kubenswrapper[4692]: E0309 09:22:22.861109 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:22 crc kubenswrapper[4692]: E0309 09:22:22.861190 4692 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.127629 4692 patch_prober.go:28] interesting pod/router-default-5444994796-d5xx5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 09 09:22:23 crc kubenswrapper[4692]: [-]has-synced failed: reason withheld Mar 09 09:22:23 crc kubenswrapper[4692]: [+]process-running ok Mar 09 09:22:23 crc kubenswrapper[4692]: healthz check failed Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.127692 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-d5xx5" podUID="cb17c405-b001-4e0a-8721-776ce83db499" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.432465 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-snq8b" Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.441222 4692 patch_prober.go:28] interesting pod/console-f9d7485db-rddqb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.441302 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-rddqb" podUID="7b8b20b2-f067-4abd-adc3-7b1a2bf36742" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.643878 4692 generic.go:334] "Generic (PLEG): container finished" podID="64c0a3f4-29f0-42fa-83bc-83e15171ed77" containerID="b861a5a13ae14d622c134d6b2ea01cde37bbfae9efc6df4d37057cf71b528c86" exitCode=0 Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.647557 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550802-w82g5" event={"ID":"64c0a3f4-29f0-42fa-83bc-83e15171ed77","Type":"ContainerDied","Data":"b861a5a13ae14d622c134d6b2ea01cde37bbfae9efc6df4d37057cf71b528c86"} Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.718308 4692 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-06 12:02:27.617017464 +0000 UTC Mar 09 09:22:23 crc kubenswrapper[4692]: I0309 09:22:23.718351 4692 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7274h40m3.898668791s for next certificate rotation Mar 09 09:22:24 crc kubenswrapper[4692]: I0309 09:22:24.127363 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:24 crc kubenswrapper[4692]: I0309 09:22:24.129503 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-d5xx5" Mar 09 09:22:24 crc kubenswrapper[4692]: I0309 09:22:24.719065 4692 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-02 09:03:22.792210175 +0000 UTC Mar 09 09:22:24 crc kubenswrapper[4692]: I0309 09:22:24.719105 4692 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6431h40m58.073107485s for next certificate rotation Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.006371 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.006442 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.006476 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.006511 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.009596 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.014638 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.015037 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.019556 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.254862 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.263294 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 09 09:22:26 crc kubenswrapper[4692]: I0309 09:22:26.272126 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:22:26 crc kubenswrapper[4692]: W0309 09:22:26.987569 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66179a0b_d877_46fb_a854_5426993f4b36.slice/crio-4892f652307c25044c45dd41ea567970ecd05cdf431f4b59fd47b39db2128ceb WatchSource:0}: Error finding container 4892f652307c25044c45dd41ea567970ecd05cdf431f4b59fd47b39db2128ceb: Status 404 returned error can't find the container with id 4892f652307c25044c45dd41ea567970ecd05cdf431f4b59fd47b39db2128ceb Mar 09 09:22:26 crc kubenswrapper[4692]: W0309 09:22:26.989030 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd840400_6e1f_4556_b60b_0cbacff31f1d.slice/crio-d5a5daa50f56a4026015ac63b025dab64e3579bf76c4b395d5a3c67048445090 WatchSource:0}: Error finding container d5a5daa50f56a4026015ac63b025dab64e3579bf76c4b395d5a3c67048445090: Status 404 returned error can't find the container with id d5a5daa50f56a4026015ac63b025dab64e3579bf76c4b395d5a3c67048445090 Mar 09 09:22:27 crc kubenswrapper[4692]: I0309 09:22:27.681079 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vbb6" event={"ID":"66179a0b-d877-46fb-a854-5426993f4b36","Type":"ContainerStarted","Data":"4892f652307c25044c45dd41ea567970ecd05cdf431f4b59fd47b39db2128ceb"} Mar 09 09:22:27 crc kubenswrapper[4692]: I0309 09:22:27.682188 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerStarted","Data":"d5a5daa50f56a4026015ac63b025dab64e3579bf76c4b395d5a3c67048445090"} Mar 09 09:22:28 crc kubenswrapper[4692]: W0309 09:22:28.109180 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d4b7447_a6e6_4530_b529_2ec461452791.slice/crio-71c712a5ba9cf979707ba7745db51f44a2867b60fb63282f872f99737e0e68c1 WatchSource:0}: Error finding container 71c712a5ba9cf979707ba7745db51f44a2867b60fb63282f872f99737e0e68c1: Status 404 returned error can't find the container with id 71c712a5ba9cf979707ba7745db51f44a2867b60fb63282f872f99737e0e68c1 Mar 09 09:22:28 crc kubenswrapper[4692]: W0309 09:22:28.112642 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15d06ca2_19f5_4c88_8c32_51e491ffa834.slice/crio-814b560c1411cc99b41a5e4bb6bb0713b89380ed4248d221aefeb3f3ce570860 WatchSource:0}: Error finding container 814b560c1411cc99b41a5e4bb6bb0713b89380ed4248d221aefeb3f3ce570860: Status 404 returned error can't find the container with id 814b560c1411cc99b41a5e4bb6bb0713b89380ed4248d221aefeb3f3ce570860 Mar 09 09:22:28 crc kubenswrapper[4692]: W0309 09:22:28.120649 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc3d45457_99aa_4e61_98f6_0fca56f42e5b.slice/crio-e0693b3265c0785cfaaf43e12a22d70d4bf57ff23ca0a7c809fadf51ed611f5c WatchSource:0}: Error finding container e0693b3265c0785cfaaf43e12a22d70d4bf57ff23ca0a7c809fadf51ed611f5c: Status 404 returned error can't find the container with id e0693b3265c0785cfaaf43e12a22d70d4bf57ff23ca0a7c809fadf51ed611f5c Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.179970 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550802-w82g5" Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.337440 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rktv7\" (UniqueName: \"kubernetes.io/projected/64c0a3f4-29f0-42fa-83bc-83e15171ed77-kube-api-access-rktv7\") pod \"64c0a3f4-29f0-42fa-83bc-83e15171ed77\" (UID: \"64c0a3f4-29f0-42fa-83bc-83e15171ed77\") " Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.354059 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c0a3f4-29f0-42fa-83bc-83e15171ed77-kube-api-access-rktv7" (OuterVolumeSpecName: "kube-api-access-rktv7") pod "64c0a3f4-29f0-42fa-83bc-83e15171ed77" (UID: "64c0a3f4-29f0-42fa-83bc-83e15171ed77"). InnerVolumeSpecName "kube-api-access-rktv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.438939 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rktv7\" (UniqueName: \"kubernetes.io/projected/64c0a3f4-29f0-42fa-83bc-83e15171ed77-kube-api-access-rktv7\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.687699 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkkhr" event={"ID":"3d4b7447-a6e6-4530-b529-2ec461452791","Type":"ContainerStarted","Data":"71c712a5ba9cf979707ba7745db51f44a2867b60fb63282f872f99737e0e68c1"} Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.688746 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerStarted","Data":"814b560c1411cc99b41a5e4bb6bb0713b89380ed4248d221aefeb3f3ce570860"} Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.689607 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ec79090-5ad5-4c78-81da-7eb7ea152e73","Type":"ContainerStarted","Data":"9ceabb42272bdd31b8046e42b687b8804f476aec912c45c5cda7332405bf2fb5"} Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.690488 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c3d45457-99aa-4e61-98f6-0fca56f42e5b","Type":"ContainerStarted","Data":"e0693b3265c0785cfaaf43e12a22d70d4bf57ff23ca0a7c809fadf51ed611f5c"} Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.691645 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550802-w82g5" event={"ID":"64c0a3f4-29f0-42fa-83bc-83e15171ed77","Type":"ContainerDied","Data":"865ca57475140c4720d8fa592829144739c7092a224cf9b0beff03dfa6ba0035"} Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.691667 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="865ca57475140c4720d8fa592829144739c7092a224cf9b0beff03dfa6ba0035" Mar 09 09:22:28 crc kubenswrapper[4692]: I0309 09:22:28.691713 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550802-w82g5" Mar 09 09:22:31 crc kubenswrapper[4692]: I0309 09:22:31.343816 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-549d7bf89b-9sxvg"] Mar 09 09:22:31 crc kubenswrapper[4692]: I0309 09:22:31.344746 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" podUID="f9b0841d-ef1d-4057-b4eb-b136af058aab" containerName="controller-manager" containerID="cri-o://4c04c043b9a5866506ea255353446b8d73f2189e127dcd71784749b9482bf9fd" gracePeriod=30 Mar 09 09:22:31 crc kubenswrapper[4692]: I0309 09:22:31.370388 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf"] Mar 09 09:22:31 crc kubenswrapper[4692]: I0309 09:22:31.370690 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerName="route-controller-manager" containerID="cri-o://81053e36c98e67cd66fb5aed96d9aafda2635780f331e2dc02d1b8c129780904" gracePeriod=30 Mar 09 09:22:32 crc kubenswrapper[4692]: I0309 09:22:32.715253 4692 generic.go:334] "Generic (PLEG): container finished" podID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerID="81053e36c98e67cd66fb5aed96d9aafda2635780f331e2dc02d1b8c129780904" exitCode=0 Mar 09 09:22:32 crc kubenswrapper[4692]: I0309 09:22:32.715383 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" event={"ID":"934d185d-4068-409d-8e31-6b8f63ca6e48","Type":"ContainerDied","Data":"81053e36c98e67cd66fb5aed96d9aafda2635780f331e2dc02d1b8c129780904"} Mar 09 09:22:32 crc kubenswrapper[4692]: I0309 09:22:32.716959 4692 generic.go:334] "Generic (PLEG): container finished" podID="f9b0841d-ef1d-4057-b4eb-b136af058aab" containerID="4c04c043b9a5866506ea255353446b8d73f2189e127dcd71784749b9482bf9fd" exitCode=0 Mar 09 09:22:32 crc kubenswrapper[4692]: I0309 09:22:32.717000 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" event={"ID":"f9b0841d-ef1d-4057-b4eb-b136af058aab","Type":"ContainerDied","Data":"4c04c043b9a5866506ea255353446b8d73f2189e127dcd71784749b9482bf9fd"} Mar 09 09:22:32 crc kubenswrapper[4692]: I0309 09:22:32.811641 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:22:32 crc kubenswrapper[4692]: E0309 09:22:32.855914 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:32 crc kubenswrapper[4692]: E0309 09:22:32.857328 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:32 crc kubenswrapper[4692]: E0309 09:22:32.858483 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:32 crc kubenswrapper[4692]: E0309 09:22:32.858526 4692 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" Mar 09 09:22:33 crc kubenswrapper[4692]: I0309 09:22:33.454677 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:22:33 crc kubenswrapper[4692]: I0309 09:22:33.458404 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-rddqb" Mar 09 09:22:33 crc kubenswrapper[4692]: I0309 09:22:33.622601 4692 patch_prober.go:28] interesting pod/controller-manager-549d7bf89b-9sxvg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" start-of-body= Mar 09 09:22:33 crc kubenswrapper[4692]: I0309 09:22:33.622671 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" podUID="f9b0841d-ef1d-4057-b4eb-b136af058aab" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" Mar 09 09:22:33 crc kubenswrapper[4692]: I0309 09:22:33.633875 4692 patch_prober.go:28] interesting pod/route-controller-manager-8f6f7fb64-8wvwf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" start-of-body= Mar 09 09:22:33 crc kubenswrapper[4692]: I0309 09:22:33.633935 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": dial tcp 10.217.0.50:8443: connect: connection refused" Mar 09 09:22:37 crc kubenswrapper[4692]: I0309 09:22:37.742834 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-c9rx4_d4063de5-2cdb-4d86-a74f-c98af0aae38d/kube-multus-additional-cni-plugins/0.log" Mar 09 09:22:37 crc kubenswrapper[4692]: I0309 09:22:37.743341 4692 generic.go:334] "Generic (PLEG): container finished" podID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" exitCode=137 Mar 09 09:22:37 crc kubenswrapper[4692]: I0309 09:22:37.743396 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" event={"ID":"d4063de5-2cdb-4d86-a74f-c98af0aae38d","Type":"ContainerDied","Data":"2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5"} Mar 09 09:22:39 crc kubenswrapper[4692]: E0309 09:22:39.346794 4692 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 09 09:22:39 crc kubenswrapper[4692]: E0309 09:22:39.346957 4692 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g6gqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4bqhb_openshift-marketplace(73819651-942b-4436-b0b1-7225cfb20ef1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 09:22:39 crc kubenswrapper[4692]: E0309 09:22:39.349560 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4bqhb" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" Mar 09 09:22:42 crc kubenswrapper[4692]: E0309 09:22:42.851638 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5 is running failed: container process not found" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:42 crc kubenswrapper[4692]: E0309 09:22:42.852844 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5 is running failed: container process not found" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:42 crc kubenswrapper[4692]: E0309 09:22:42.853111 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5 is running failed: container process not found" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 09 09:22:42 crc kubenswrapper[4692]: E0309 09:22:42.853146 4692 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" Mar 09 09:22:43 crc kubenswrapper[4692]: I0309 09:22:43.085081 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 09 09:22:43 crc kubenswrapper[4692]: I0309 09:22:43.787653 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-4gfv7" Mar 09 09:22:43 crc kubenswrapper[4692]: I0309 09:22:43.809726 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=0.809706254 podStartE2EDuration="809.706254ms" podCreationTimestamp="2026-03-09 09:22:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:43.807267849 +0000 UTC m=+164.632003440" watchObservedRunningTime="2026-03-09 09:22:43.809706254 +0000 UTC m=+164.634441865" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.085629 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.623021 4692 patch_prober.go:28] interesting pod/controller-manager-549d7bf89b-9sxvg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: i/o timeout" start-of-body= Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.623083 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" podUID="f9b0841d-ef1d-4057-b4eb-b136af058aab" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: i/o timeout" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.633949 4692 patch_prober.go:28] interesting pod/route-controller-manager-8f6f7fb64-8wvwf container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.50:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.633992 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.50:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.909657 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 09 09:22:44 crc kubenswrapper[4692]: E0309 09:22:44.910124 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c0a3f4-29f0-42fa-83bc-83e15171ed77" containerName="oc" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.910232 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c0a3f4-29f0-42fa-83bc-83e15171ed77" containerName="oc" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.910479 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c0a3f4-29f0-42fa-83bc-83e15171ed77" containerName="oc" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.911328 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.918706 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.949338 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=0.949319638 podStartE2EDuration="949.319638ms" podCreationTimestamp="2026-03-09 09:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:44.946484812 +0000 UTC m=+165.771220403" watchObservedRunningTime="2026-03-09 09:22:44.949319638 +0000 UTC m=+165.774055219" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.968546 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0128b73b-793f-454e-a1c6-250f99c5a748-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:44 crc kubenswrapper[4692]: I0309 09:22:44.968632 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0128b73b-793f-454e-a1c6-250f99c5a748-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.069670 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0128b73b-793f-454e-a1c6-250f99c5a748-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.069764 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0128b73b-793f-454e-a1c6-250f99c5a748-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.069832 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0128b73b-793f-454e-a1c6-250f99c5a748-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.180587 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0128b73b-793f-454e-a1c6-250f99c5a748-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:45 crc kubenswrapper[4692]: E0309 09:22:45.200747 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4bqhb" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.230637 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.271032 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.299944 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4"] Mar 09 09:22:45 crc kubenswrapper[4692]: E0309 09:22:45.300226 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerName="route-controller-manager" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.300242 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerName="route-controller-manager" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.300372 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" containerName="route-controller-manager" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.300827 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.306867 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4"] Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.372580 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/934d185d-4068-409d-8e31-6b8f63ca6e48-serving-cert\") pod \"934d185d-4068-409d-8e31-6b8f63ca6e48\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.372623 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-config\") pod \"934d185d-4068-409d-8e31-6b8f63ca6e48\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.372647 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-client-ca\") pod \"934d185d-4068-409d-8e31-6b8f63ca6e48\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.372715 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-752bq\" (UniqueName: \"kubernetes.io/projected/934d185d-4068-409d-8e31-6b8f63ca6e48-kube-api-access-752bq\") pod \"934d185d-4068-409d-8e31-6b8f63ca6e48\" (UID: \"934d185d-4068-409d-8e31-6b8f63ca6e48\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.374567 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-config" (OuterVolumeSpecName: "config") pod "934d185d-4068-409d-8e31-6b8f63ca6e48" (UID: "934d185d-4068-409d-8e31-6b8f63ca6e48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.377290 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-client-ca" (OuterVolumeSpecName: "client-ca") pod "934d185d-4068-409d-8e31-6b8f63ca6e48" (UID: "934d185d-4068-409d-8e31-6b8f63ca6e48"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.377655 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.474699 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-config\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.474844 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b2jr\" (UniqueName: \"kubernetes.io/projected/39561b1e-95e6-4c18-83dd-09930060d986-kube-api-access-9b2jr\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.474916 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-client-ca\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.474960 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39561b1e-95e6-4c18-83dd-09930060d986-serving-cert\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.475012 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.475041 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/934d185d-4068-409d-8e31-6b8f63ca6e48-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576242 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-config\") pod \"f9b0841d-ef1d-4057-b4eb-b136af058aab\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576324 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-proxy-ca-bundles\") pod \"f9b0841d-ef1d-4057-b4eb-b136af058aab\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576368 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2ldl\" (UniqueName: \"kubernetes.io/projected/f9b0841d-ef1d-4057-b4eb-b136af058aab-kube-api-access-j2ldl\") pod \"f9b0841d-ef1d-4057-b4eb-b136af058aab\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576389 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9b0841d-ef1d-4057-b4eb-b136af058aab-serving-cert\") pod \"f9b0841d-ef1d-4057-b4eb-b136af058aab\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576412 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-client-ca\") pod \"f9b0841d-ef1d-4057-b4eb-b136af058aab\" (UID: \"f9b0841d-ef1d-4057-b4eb-b136af058aab\") " Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576481 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39561b1e-95e6-4c18-83dd-09930060d986-serving-cert\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576517 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-config\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576568 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b2jr\" (UniqueName: \"kubernetes.io/projected/39561b1e-95e6-4c18-83dd-09930060d986-kube-api-access-9b2jr\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.576603 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-client-ca\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.577249 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f9b0841d-ef1d-4057-b4eb-b136af058aab" (UID: "f9b0841d-ef1d-4057-b4eb-b136af058aab"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.577296 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-config" (OuterVolumeSpecName: "config") pod "f9b0841d-ef1d-4057-b4eb-b136af058aab" (UID: "f9b0841d-ef1d-4057-b4eb-b136af058aab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.577961 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-client-ca" (OuterVolumeSpecName: "client-ca") pod "f9b0841d-ef1d-4057-b4eb-b136af058aab" (UID: "f9b0841d-ef1d-4057-b4eb-b136af058aab"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.579928 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-client-ca\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.581385 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39561b1e-95e6-4c18-83dd-09930060d986-serving-cert\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.581638 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-config\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.592615 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b2jr\" (UniqueName: \"kubernetes.io/projected/39561b1e-95e6-4c18-83dd-09930060d986-kube-api-access-9b2jr\") pod \"route-controller-manager-7f5ccd5b88-b7ps4\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.678113 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.678478 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.678498 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9b0841d-ef1d-4057-b4eb-b136af058aab-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.739444 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 09 09:22:45 crc kubenswrapper[4692]: W0309 09:22:45.743748 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod0128b73b_793f_454e_a1c6_250f99c5a748.slice/crio-61c1084ca743318ea28426049ffe3051f21b80381ce3446d16d29570f231bc85 WatchSource:0}: Error finding container 61c1084ca743318ea28426049ffe3051f21b80381ce3446d16d29570f231bc85: Status 404 returned error can't find the container with id 61c1084ca743318ea28426049ffe3051f21b80381ce3446d16d29570f231bc85 Mar 09 09:22:45 crc kubenswrapper[4692]: W0309 09:22:45.746736 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-927763fd507331f08a5d98214b053dd8d33f41d5083a24422c0cf9269eb1a131 WatchSource:0}: Error finding container 927763fd507331f08a5d98214b053dd8d33f41d5083a24422c0cf9269eb1a131: Status 404 returned error can't find the container with id 927763fd507331f08a5d98214b053dd8d33f41d5083a24422c0cf9269eb1a131 Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.760525 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.784421 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"c0b3c1adbc4226bc3c950e7211f4dda3142c0f4f65472244d1bbc44cbbf9c37f"} Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.786160 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" event={"ID":"934d185d-4068-409d-8e31-6b8f63ca6e48","Type":"ContainerDied","Data":"56cd50398bb4dee57f9f5c41d9258c219a85a789e4aca7fb16193cfc132a0dbb"} Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.786249 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.786254 4692 scope.go:117] "RemoveContainer" containerID="81053e36c98e67cd66fb5aed96d9aafda2635780f331e2dc02d1b8c129780904" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.787577 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0128b73b-793f-454e-a1c6-250f99c5a748","Type":"ContainerStarted","Data":"61c1084ca743318ea28426049ffe3051f21b80381ce3446d16d29570f231bc85"} Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.789077 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"927763fd507331f08a5d98214b053dd8d33f41d5083a24422c0cf9269eb1a131"} Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.790908 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c75a4a5ad003a9c1c760cf093e9e23cdb77f9a8691bf985bcb75fa7d32dd95f0"} Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.792330 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" event={"ID":"f9b0841d-ef1d-4057-b4eb-b136af058aab","Type":"ContainerDied","Data":"a84f1cfbe073fe80c0aa6f75d63b0a2119821d39be5840af9486a37c4e74499c"} Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.792462 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-549d7bf89b-9sxvg" Mar 09 09:22:45 crc kubenswrapper[4692]: I0309 09:22:45.802806 4692 scope.go:117] "RemoveContainer" containerID="4c04c043b9a5866506ea255353446b8d73f2189e127dcd71784749b9482bf9fd" Mar 09 09:22:46 crc kubenswrapper[4692]: I0309 09:22:46.142841 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4"] Mar 09 09:22:46 crc kubenswrapper[4692]: W0309 09:22:46.442132 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39561b1e_95e6_4c18_83dd_09930060d986.slice/crio-7ad71de5e773233b3f73cff196739f67f5e033c493ec6e84a773df37efaf3231 WatchSource:0}: Error finding container 7ad71de5e773233b3f73cff196739f67f5e033c493ec6e84a773df37efaf3231: Status 404 returned error can't find the container with id 7ad71de5e773233b3f73cff196739f67f5e033c493ec6e84a773df37efaf3231 Mar 09 09:22:46 crc kubenswrapper[4692]: I0309 09:22:46.799493 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" event={"ID":"39561b1e-95e6-4c18-83dd-09930060d986","Type":"ContainerStarted","Data":"7ad71de5e773233b3f73cff196739f67f5e033c493ec6e84a773df37efaf3231"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.316120 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-84fff57f76-rsxjg"] Mar 09 09:22:47 crc kubenswrapper[4692]: E0309 09:22:47.316968 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b0841d-ef1d-4057-b4eb-b136af058aab" containerName="controller-manager" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.317009 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b0841d-ef1d-4057-b4eb-b136af058aab" containerName="controller-manager" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.317264 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9b0841d-ef1d-4057-b4eb-b136af058aab" containerName="controller-manager" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.317958 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.330422 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-84fff57f76-rsxjg"] Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.397538 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf8nx\" (UniqueName: \"kubernetes.io/projected/005399a7-db26-44d0-a0fc-bce24138eb7b-kube-api-access-kf8nx\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.397620 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-proxy-ca-bundles\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.397659 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-client-ca\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.397684 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-config\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.397719 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005399a7-db26-44d0-a0fc-bce24138eb7b-serving-cert\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.499075 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005399a7-db26-44d0-a0fc-bce24138eb7b-serving-cert\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.500455 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf8nx\" (UniqueName: \"kubernetes.io/projected/005399a7-db26-44d0-a0fc-bce24138eb7b-kube-api-access-kf8nx\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.500522 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-proxy-ca-bundles\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.500553 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-client-ca\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.500579 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-config\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.502381 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-config\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.502585 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-proxy-ca-bundles\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.502796 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-client-ca\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.509485 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005399a7-db26-44d0-a0fc-bce24138eb7b-serving-cert\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.523065 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf8nx\" (UniqueName: \"kubernetes.io/projected/005399a7-db26-44d0-a0fc-bce24138eb7b-kube-api-access-kf8nx\") pod \"controller-manager-84fff57f76-rsxjg\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.637573 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:47 crc kubenswrapper[4692]: E0309 09:22:47.785748 4692 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 09 09:22:47 crc kubenswrapper[4692]: E0309 09:22:47.785980 4692 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w9w5h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-q9nxd_openshift-marketplace(27500a21-baed-4a7c-9af8-e616fee71857): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 09:22:47 crc kubenswrapper[4692]: E0309 09:22:47.787749 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-q9nxd" podUID="27500a21-baed-4a7c-9af8-e616fee71857" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.809857 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkkhr" event={"ID":"3d4b7447-a6e6-4530-b529-2ec461452791","Type":"ContainerStarted","Data":"08664efca426d1881961484a5961fc666e14026ba62b4d2653ee4e236c61172c"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.812365 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0128b73b-793f-454e-a1c6-250f99c5a748","Type":"ContainerStarted","Data":"f8a47c1e6684a9b4eee1980ee36de88a31999ad0a04850198aef1126ee29a505"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.814268 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vbb6" event={"ID":"66179a0b-d877-46fb-a854-5426993f4b36","Type":"ContainerStarted","Data":"d02ca25939ba28102a59713650f778d6e75c59d85875075436a4ffd69ee7c0c4"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.817790 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerStarted","Data":"4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.819324 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ec79090-5ad5-4c78-81da-7eb7ea152e73","Type":"ContainerStarted","Data":"387b4aba65d8bf78f7e5793533a1c4f484e361f9da3bbd55d7109cc56b037569"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.825177 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" event={"ID":"39561b1e-95e6-4c18-83dd-09930060d986","Type":"ContainerStarted","Data":"1d8f47335ea3090f1f325d8f1dfd761d242abed8068a348c479b82c0b005b558"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.828558 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c3d45457-99aa-4e61-98f6-0fca56f42e5b","Type":"ContainerStarted","Data":"0750b1eec5b09171e4fa44adbcd30bde0380811eb1572ec9b9cf8cc115d6ac82"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.830148 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"efbf0a3deb7f347f41630260a749d9f25ba0e7d5ce8a2b7974856ce86ccd0a1e"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.833651 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"597e9f9c5dfd6a7dfe233a88c4d3849bc303675b41a2887bd406d613e9f22736"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.834551 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.840838 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerStarted","Data":"104b76fb6f9d88f766ca6910007c7dc64151f713d824dfbc8ff9a211a91cb5ce"} Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.841647 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=34.841625575 podStartE2EDuration="34.841625575s" podCreationTimestamp="2026-03-09 09:22:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:47.839211381 +0000 UTC m=+168.663946962" watchObservedRunningTime="2026-03-09 09:22:47.841625575 +0000 UTC m=+168.666361156" Mar 09 09:22:47 crc kubenswrapper[4692]: I0309 09:22:47.845048 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"3e597cb74ba5336e2b6bece0fc29400f91bbd23b8ee07a7605636c69409349fa"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.253012 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9b0841d-ef1d-4057-b4eb-b136af058aab-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f9b0841d-ef1d-4057-b4eb-b136af058aab" (UID: "f9b0841d-ef1d-4057-b4eb-b136af058aab"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.253234 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/934d185d-4068-409d-8e31-6b8f63ca6e48-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "934d185d-4068-409d-8e31-6b8f63ca6e48" (UID: "934d185d-4068-409d-8e31-6b8f63ca6e48"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.253528 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9b0841d-ef1d-4057-b4eb-b136af058aab-kube-api-access-j2ldl" (OuterVolumeSpecName: "kube-api-access-j2ldl") pod "f9b0841d-ef1d-4057-b4eb-b136af058aab" (UID: "f9b0841d-ef1d-4057-b4eb-b136af058aab"). InnerVolumeSpecName "kube-api-access-j2ldl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.253581 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/934d185d-4068-409d-8e31-6b8f63ca6e48-kube-api-access-752bq" (OuterVolumeSpecName: "kube-api-access-752bq") pod "934d185d-4068-409d-8e31-6b8f63ca6e48" (UID: "934d185d-4068-409d-8e31-6b8f63ca6e48"). InnerVolumeSpecName "kube-api-access-752bq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.257028 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-q9nxd" podUID="27500a21-baed-4a7c-9af8-e616fee71857" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.313072 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/934d185d-4068-409d-8e31-6b8f63ca6e48-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.313450 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2ldl\" (UniqueName: \"kubernetes.io/projected/f9b0841d-ef1d-4057-b4eb-b136af058aab-kube-api-access-j2ldl\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.313467 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9b0841d-ef1d-4057-b4eb-b136af058aab-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.313479 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-752bq\" (UniqueName: \"kubernetes.io/projected/934d185d-4068-409d-8e31-6b8f63ca6e48-kube-api-access-752bq\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.373846 4692 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.374023 4692 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jgtzd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qgmsd_openshift-marketplace(2f27156a-57b6-4230-aa9b-1809cbfb3569): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.375192 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qgmsd" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.471854 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-84fff57f76-rsxjg"] Mar 09 09:22:48 crc kubenswrapper[4692]: W0309 09:22:48.513846 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod005399a7_db26_44d0_a0fc_bce24138eb7b.slice/crio-aed99806e2b895cbd1f3c5d7889395bbca314692d09da289bd922a7113bacf1b WatchSource:0}: Error finding container aed99806e2b895cbd1f3c5d7889395bbca314692d09da289bd922a7113bacf1b: Status 404 returned error can't find the container with id aed99806e2b895cbd1f3c5d7889395bbca314692d09da289bd922a7113bacf1b Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.612754 4692 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.612932 4692 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6gkcn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ktmgp_openshift-marketplace(067264d3-8fc2-46cd-8d94-e89f9e25a500): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.614254 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ktmgp" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.847086 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-c9rx4_d4063de5-2cdb-4d86-a74f-c98af0aae38d/kube-multus-additional-cni-plugins/0.log" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.847176 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.851412 4692 generic.go:334] "Generic (PLEG): container finished" podID="4ec79090-5ad5-4c78-81da-7eb7ea152e73" containerID="387b4aba65d8bf78f7e5793533a1c4f484e361f9da3bbd55d7109cc56b037569" exitCode=0 Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.851494 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ec79090-5ad5-4c78-81da-7eb7ea152e73","Type":"ContainerDied","Data":"387b4aba65d8bf78f7e5793533a1c4f484e361f9da3bbd55d7109cc56b037569"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.853186 4692 generic.go:334] "Generic (PLEG): container finished" podID="3d4b7447-a6e6-4530-b529-2ec461452791" containerID="08664efca426d1881961484a5961fc666e14026ba62b4d2653ee4e236c61172c" exitCode=0 Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.853217 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkkhr" event={"ID":"3d4b7447-a6e6-4530-b529-2ec461452791","Type":"ContainerDied","Data":"08664efca426d1881961484a5961fc666e14026ba62b4d2653ee4e236c61172c"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.854647 4692 generic.go:334] "Generic (PLEG): container finished" podID="0128b73b-793f-454e-a1c6-250f99c5a748" containerID="f8a47c1e6684a9b4eee1980ee36de88a31999ad0a04850198aef1126ee29a505" exitCode=0 Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.854860 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0128b73b-793f-454e-a1c6-250f99c5a748","Type":"ContainerDied","Data":"f8a47c1e6684a9b4eee1980ee36de88a31999ad0a04850198aef1126ee29a505"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.857289 4692 generic.go:334] "Generic (PLEG): container finished" podID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerID="4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba" exitCode=0 Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.857324 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerDied","Data":"4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.858683 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-c9rx4_d4063de5-2cdb-4d86-a74f-c98af0aae38d/kube-multus-additional-cni-plugins/0.log" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.858741 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" event={"ID":"d4063de5-2cdb-4d86-a74f-c98af0aae38d","Type":"ContainerDied","Data":"a9341bbcca19f0316c60526aa64c3e5761de4566f421911a4e09c5249e1143f3"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.858827 4692 scope.go:117] "RemoveContainer" containerID="2e4c190f5be6c397745d938075359d8d5498cc2f9bd077b430aabb0e031be9e5" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.858926 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-c9rx4" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.870883 4692 generic.go:334] "Generic (PLEG): container finished" podID="c3d45457-99aa-4e61-98f6-0fca56f42e5b" containerID="0750b1eec5b09171e4fa44adbcd30bde0380811eb1572ec9b9cf8cc115d6ac82" exitCode=0 Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.870975 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c3d45457-99aa-4e61-98f6-0fca56f42e5b","Type":"ContainerDied","Data":"0750b1eec5b09171e4fa44adbcd30bde0380811eb1572ec9b9cf8cc115d6ac82"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.874460 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" event={"ID":"005399a7-db26-44d0-a0fc-bce24138eb7b","Type":"ContainerStarted","Data":"aed99806e2b895cbd1f3c5d7889395bbca314692d09da289bd922a7113bacf1b"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.876041 4692 generic.go:334] "Generic (PLEG): container finished" podID="66179a0b-d877-46fb-a854-5426993f4b36" containerID="d02ca25939ba28102a59713650f778d6e75c59d85875075436a4ffd69ee7c0c4" exitCode=0 Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.876102 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vbb6" event={"ID":"66179a0b-d877-46fb-a854-5426993f4b36","Type":"ContainerDied","Data":"d02ca25939ba28102a59713650f778d6e75c59d85875075436a4ffd69ee7c0c4"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.881487 4692 generic.go:334] "Generic (PLEG): container finished" podID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerID="104b76fb6f9d88f766ca6910007c7dc64151f713d824dfbc8ff9a211a91cb5ce" exitCode=0 Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.881533 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerDied","Data":"104b76fb6f9d88f766ca6910007c7dc64151f713d824dfbc8ff9a211a91cb5ce"} Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.882981 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.883813 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ktmgp" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" Mar 09 09:22:48 crc kubenswrapper[4692]: E0309 09:22:48.885976 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qgmsd" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" Mar 09 09:22:48 crc kubenswrapper[4692]: I0309 09:22:48.892900 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.012534 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-549d7bf89b-9sxvg"] Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.017597 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-549d7bf89b-9sxvg"] Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.028916 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d4063de5-2cdb-4d86-a74f-c98af0aae38d-cni-sysctl-allowlist\") pod \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.029012 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/d4063de5-2cdb-4d86-a74f-c98af0aae38d-ready\") pod \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.029051 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4063de5-2cdb-4d86-a74f-c98af0aae38d-tuning-conf-dir\") pod \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.029153 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkb9w\" (UniqueName: \"kubernetes.io/projected/d4063de5-2cdb-4d86-a74f-c98af0aae38d-kube-api-access-rkb9w\") pod \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\" (UID: \"d4063de5-2cdb-4d86-a74f-c98af0aae38d\") " Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.030778 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4063de5-2cdb-4d86-a74f-c98af0aae38d-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "d4063de5-2cdb-4d86-a74f-c98af0aae38d" (UID: "d4063de5-2cdb-4d86-a74f-c98af0aae38d"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.031233 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4063de5-2cdb-4d86-a74f-c98af0aae38d-ready" (OuterVolumeSpecName: "ready") pod "d4063de5-2cdb-4d86-a74f-c98af0aae38d" (UID: "d4063de5-2cdb-4d86-a74f-c98af0aae38d"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.031470 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4063de5-2cdb-4d86-a74f-c98af0aae38d-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "d4063de5-2cdb-4d86-a74f-c98af0aae38d" (UID: "d4063de5-2cdb-4d86-a74f-c98af0aae38d"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.041346 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4063de5-2cdb-4d86-a74f-c98af0aae38d-kube-api-access-rkb9w" (OuterVolumeSpecName: "kube-api-access-rkb9w") pod "d4063de5-2cdb-4d86-a74f-c98af0aae38d" (UID: "d4063de5-2cdb-4d86-a74f-c98af0aae38d"). InnerVolumeSpecName "kube-api-access-rkb9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.083756 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" podStartSLOduration=18.083734474 podStartE2EDuration="18.083734474s" podCreationTimestamp="2026-03-09 09:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:49.054541751 +0000 UTC m=+169.879277342" watchObservedRunningTime="2026-03-09 09:22:49.083734474 +0000 UTC m=+169.908470065" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.137447 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf"] Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.137508 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8f6f7fb64-8wvwf"] Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.137868 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkb9w\" (UniqueName: \"kubernetes.io/projected/d4063de5-2cdb-4d86-a74f-c98af0aae38d-kube-api-access-rkb9w\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.137891 4692 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d4063de5-2cdb-4d86-a74f-c98af0aae38d-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.137903 4692 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/d4063de5-2cdb-4d86-a74f-c98af0aae38d-ready\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.137913 4692 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d4063de5-2cdb-4d86-a74f-c98af0aae38d-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.189061 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-c9rx4"] Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.193253 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-c9rx4"] Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.920896 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" event={"ID":"005399a7-db26-44d0-a0fc-bce24138eb7b","Type":"ContainerStarted","Data":"c63066a94b2ada58f76dceecfbc496d3a72fbc538542857d22ba79c17cd4c951"} Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.921737 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.934777 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:49 crc kubenswrapper[4692]: I0309 09:22:49.980237 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" podStartSLOduration=18.980213412 podStartE2EDuration="18.980213412s" podCreationTimestamp="2026-03-09 09:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:49.964857992 +0000 UTC m=+170.789593583" watchObservedRunningTime="2026-03-09 09:22:49.980213412 +0000 UTC m=+170.804948993" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.081438 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="934d185d-4068-409d-8e31-6b8f63ca6e48" path="/var/lib/kubelet/pods/934d185d-4068-409d-8e31-6b8f63ca6e48/volumes" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.082234 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" path="/var/lib/kubelet/pods/d4063de5-2cdb-4d86-a74f-c98af0aae38d/volumes" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.085066 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9b0841d-ef1d-4057-b4eb-b136af058aab" path="/var/lib/kubelet/pods/f9b0841d-ef1d-4057-b4eb-b136af058aab/volumes" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.256254 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.331374 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.334861 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.355831 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kubelet-dir\") pod \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.356086 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4ec79090-5ad5-4c78-81da-7eb7ea152e73" (UID: "4ec79090-5ad5-4c78-81da-7eb7ea152e73"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.356124 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kube-api-access\") pod \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\" (UID: \"4ec79090-5ad5-4c78-81da-7eb7ea152e73\") " Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.356598 4692 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.365614 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4ec79090-5ad5-4c78-81da-7eb7ea152e73" (UID: "4ec79090-5ad5-4c78-81da-7eb7ea152e73"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457349 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kubelet-dir\") pod \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457434 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0128b73b-793f-454e-a1c6-250f99c5a748-kube-api-access\") pod \"0128b73b-793f-454e-a1c6-250f99c5a748\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457528 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kube-api-access\") pod \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\" (UID: \"c3d45457-99aa-4e61-98f6-0fca56f42e5b\") " Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457576 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0128b73b-793f-454e-a1c6-250f99c5a748-kubelet-dir\") pod \"0128b73b-793f-454e-a1c6-250f99c5a748\" (UID: \"0128b73b-793f-454e-a1c6-250f99c5a748\") " Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457536 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c3d45457-99aa-4e61-98f6-0fca56f42e5b" (UID: "c3d45457-99aa-4e61-98f6-0fca56f42e5b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457867 4692 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457890 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ec79090-5ad5-4c78-81da-7eb7ea152e73-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.457934 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0128b73b-793f-454e-a1c6-250f99c5a748-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0128b73b-793f-454e-a1c6-250f99c5a748" (UID: "0128b73b-793f-454e-a1c6-250f99c5a748"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.461898 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c3d45457-99aa-4e61-98f6-0fca56f42e5b" (UID: "c3d45457-99aa-4e61-98f6-0fca56f42e5b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.461951 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0128b73b-793f-454e-a1c6-250f99c5a748-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0128b73b-793f-454e-a1c6-250f99c5a748" (UID: "0128b73b-793f-454e-a1c6-250f99c5a748"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.558759 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0128b73b-793f-454e-a1c6-250f99c5a748-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.558804 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c3d45457-99aa-4e61-98f6-0fca56f42e5b-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.558829 4692 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0128b73b-793f-454e-a1c6-250f99c5a748-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.930692 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"4ec79090-5ad5-4c78-81da-7eb7ea152e73","Type":"ContainerDied","Data":"9ceabb42272bdd31b8046e42b687b8804f476aec912c45c5cda7332405bf2fb5"} Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.930723 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.930745 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ceabb42272bdd31b8046e42b687b8804f476aec912c45c5cda7332405bf2fb5" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.933146 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c3d45457-99aa-4e61-98f6-0fca56f42e5b","Type":"ContainerDied","Data":"e0693b3265c0785cfaaf43e12a22d70d4bf57ff23ca0a7c809fadf51ed611f5c"} Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.933195 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0693b3265c0785cfaaf43e12a22d70d4bf57ff23ca0a7c809fadf51ed611f5c" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.933173 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.934788 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0128b73b-793f-454e-a1c6-250f99c5a748","Type":"ContainerDied","Data":"61c1084ca743318ea28426049ffe3051f21b80381ce3446d16d29570f231bc85"} Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.934866 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 09 09:22:50 crc kubenswrapper[4692]: I0309 09:22:50.934868 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61c1084ca743318ea28426049ffe3051f21b80381ce3446d16d29570f231bc85" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.101322 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 09 09:22:51 crc kubenswrapper[4692]: E0309 09:22:51.103445 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec79090-5ad5-4c78-81da-7eb7ea152e73" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103469 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec79090-5ad5-4c78-81da-7eb7ea152e73" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: E0309 09:22:51.103484 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103492 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" Mar 09 09:22:51 crc kubenswrapper[4692]: E0309 09:22:51.103503 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0128b73b-793f-454e-a1c6-250f99c5a748" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103512 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="0128b73b-793f-454e-a1c6-250f99c5a748" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: E0309 09:22:51.103525 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3d45457-99aa-4e61-98f6-0fca56f42e5b" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103534 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3d45457-99aa-4e61-98f6-0fca56f42e5b" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103662 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3d45457-99aa-4e61-98f6-0fca56f42e5b" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103680 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4063de5-2cdb-4d86-a74f-c98af0aae38d" containerName="kube-multus-additional-cni-plugins" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103692 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec79090-5ad5-4c78-81da-7eb7ea152e73" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.103701 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="0128b73b-793f-454e-a1c6-250f99c5a748" containerName="pruner" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.104274 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.108613 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.108654 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.110485 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.266362 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.266516 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-var-lock\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.266537 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kube-api-access\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.344501 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-84fff57f76-rsxjg"] Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.357555 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4"] Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.367818 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-var-lock\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.367862 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kube-api-access\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.367896 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.368025 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.368061 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-var-lock\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.387441 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kube-api-access\") pod \"installer-9-crc\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.439341 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.888471 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 09 09:22:51 crc kubenswrapper[4692]: I0309 09:22:51.945649 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" podUID="39561b1e-95e6-4c18-83dd-09930060d986" containerName="route-controller-manager" containerID="cri-o://1d8f47335ea3090f1f325d8f1dfd761d242abed8068a348c479b82c0b005b558" gracePeriod=30 Mar 09 09:22:52 crc kubenswrapper[4692]: I0309 09:22:52.953971 4692 generic.go:334] "Generic (PLEG): container finished" podID="39561b1e-95e6-4c18-83dd-09930060d986" containerID="1d8f47335ea3090f1f325d8f1dfd761d242abed8068a348c479b82c0b005b558" exitCode=0 Mar 09 09:22:52 crc kubenswrapper[4692]: I0309 09:22:52.954026 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" event={"ID":"39561b1e-95e6-4c18-83dd-09930060d986","Type":"ContainerDied","Data":"1d8f47335ea3090f1f325d8f1dfd761d242abed8068a348c479b82c0b005b558"} Mar 09 09:22:52 crc kubenswrapper[4692]: I0309 09:22:52.954196 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" podUID="005399a7-db26-44d0-a0fc-bce24138eb7b" containerName="controller-manager" containerID="cri-o://c63066a94b2ada58f76dceecfbc496d3a72fbc538542857d22ba79c17cd4c951" gracePeriod=30 Mar 09 09:22:53 crc kubenswrapper[4692]: W0309 09:22:53.771398 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc61e177f_a1d1_4a8b_a7d7_24035a88f54e.slice/crio-6304cbe8a68ce7c136b8c307f75c2e7a791cd2c35d9a065a35fc41ade603a186 WatchSource:0}: Error finding container 6304cbe8a68ce7c136b8c307f75c2e7a791cd2c35d9a065a35fc41ade603a186: Status 404 returned error can't find the container with id 6304cbe8a68ce7c136b8c307f75c2e7a791cd2c35d9a065a35fc41ade603a186 Mar 09 09:22:53 crc kubenswrapper[4692]: I0309 09:22:53.974442 4692 generic.go:334] "Generic (PLEG): container finished" podID="005399a7-db26-44d0-a0fc-bce24138eb7b" containerID="c63066a94b2ada58f76dceecfbc496d3a72fbc538542857d22ba79c17cd4c951" exitCode=0 Mar 09 09:22:53 crc kubenswrapper[4692]: I0309 09:22:53.974564 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" event={"ID":"005399a7-db26-44d0-a0fc-bce24138eb7b","Type":"ContainerDied","Data":"c63066a94b2ada58f76dceecfbc496d3a72fbc538542857d22ba79c17cd4c951"} Mar 09 09:22:53 crc kubenswrapper[4692]: I0309 09:22:53.984145 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c61e177f-a1d1-4a8b-a7d7-24035a88f54e","Type":"ContainerStarted","Data":"6304cbe8a68ce7c136b8c307f75c2e7a791cd2c35d9a065a35fc41ade603a186"} Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.167017 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.173931 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.223269 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk"] Mar 09 09:22:54 crc kubenswrapper[4692]: E0309 09:22:54.223540 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39561b1e-95e6-4c18-83dd-09930060d986" containerName="route-controller-manager" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.223563 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="39561b1e-95e6-4c18-83dd-09930060d986" containerName="route-controller-manager" Mar 09 09:22:54 crc kubenswrapper[4692]: E0309 09:22:54.223574 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="005399a7-db26-44d0-a0fc-bce24138eb7b" containerName="controller-manager" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.223582 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="005399a7-db26-44d0-a0fc-bce24138eb7b" containerName="controller-manager" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.223723 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="39561b1e-95e6-4c18-83dd-09930060d986" containerName="route-controller-manager" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.223737 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="005399a7-db26-44d0-a0fc-bce24138eb7b" containerName="controller-manager" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.224271 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.232202 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk"] Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.311795 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39561b1e-95e6-4c18-83dd-09930060d986-serving-cert\") pod \"39561b1e-95e6-4c18-83dd-09930060d986\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.311847 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-config\") pod \"005399a7-db26-44d0-a0fc-bce24138eb7b\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.311888 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-proxy-ca-bundles\") pod \"005399a7-db26-44d0-a0fc-bce24138eb7b\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.311910 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-config\") pod \"39561b1e-95e6-4c18-83dd-09930060d986\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.311945 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-client-ca\") pod \"39561b1e-95e6-4c18-83dd-09930060d986\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.311969 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-client-ca\") pod \"005399a7-db26-44d0-a0fc-bce24138eb7b\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.312009 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf8nx\" (UniqueName: \"kubernetes.io/projected/005399a7-db26-44d0-a0fc-bce24138eb7b-kube-api-access-kf8nx\") pod \"005399a7-db26-44d0-a0fc-bce24138eb7b\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.312318 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005399a7-db26-44d0-a0fc-bce24138eb7b-serving-cert\") pod \"005399a7-db26-44d0-a0fc-bce24138eb7b\" (UID: \"005399a7-db26-44d0-a0fc-bce24138eb7b\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.312393 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b2jr\" (UniqueName: \"kubernetes.io/projected/39561b1e-95e6-4c18-83dd-09930060d986-kube-api-access-9b2jr\") pod \"39561b1e-95e6-4c18-83dd-09930060d986\" (UID: \"39561b1e-95e6-4c18-83dd-09930060d986\") " Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.314471 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "005399a7-db26-44d0-a0fc-bce24138eb7b" (UID: "005399a7-db26-44d0-a0fc-bce24138eb7b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.314773 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-client-ca" (OuterVolumeSpecName: "client-ca") pod "005399a7-db26-44d0-a0fc-bce24138eb7b" (UID: "005399a7-db26-44d0-a0fc-bce24138eb7b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.315783 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-config" (OuterVolumeSpecName: "config") pod "39561b1e-95e6-4c18-83dd-09930060d986" (UID: "39561b1e-95e6-4c18-83dd-09930060d986"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.317288 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.317330 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.317345 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.317366 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-config" (OuterVolumeSpecName: "config") pod "005399a7-db26-44d0-a0fc-bce24138eb7b" (UID: "005399a7-db26-44d0-a0fc-bce24138eb7b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.318336 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-client-ca" (OuterVolumeSpecName: "client-ca") pod "39561b1e-95e6-4c18-83dd-09930060d986" (UID: "39561b1e-95e6-4c18-83dd-09930060d986"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.320032 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/005399a7-db26-44d0-a0fc-bce24138eb7b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "005399a7-db26-44d0-a0fc-bce24138eb7b" (UID: "005399a7-db26-44d0-a0fc-bce24138eb7b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.320473 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/005399a7-db26-44d0-a0fc-bce24138eb7b-kube-api-access-kf8nx" (OuterVolumeSpecName: "kube-api-access-kf8nx") pod "005399a7-db26-44d0-a0fc-bce24138eb7b" (UID: "005399a7-db26-44d0-a0fc-bce24138eb7b"). InnerVolumeSpecName "kube-api-access-kf8nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.329068 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39561b1e-95e6-4c18-83dd-09930060d986-kube-api-access-9b2jr" (OuterVolumeSpecName: "kube-api-access-9b2jr") pod "39561b1e-95e6-4c18-83dd-09930060d986" (UID: "39561b1e-95e6-4c18-83dd-09930060d986"). InnerVolumeSpecName "kube-api-access-9b2jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.334585 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39561b1e-95e6-4c18-83dd-09930060d986-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "39561b1e-95e6-4c18-83dd-09930060d986" (UID: "39561b1e-95e6-4c18-83dd-09930060d986"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418176 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-config\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418257 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa3427b-4c78-404f-9cee-2f065504baec-serving-cert\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418307 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plvrs\" (UniqueName: \"kubernetes.io/projected/6aa3427b-4c78-404f-9cee-2f065504baec-kube-api-access-plvrs\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418367 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-client-ca\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418422 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b2jr\" (UniqueName: \"kubernetes.io/projected/39561b1e-95e6-4c18-83dd-09930060d986-kube-api-access-9b2jr\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418434 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39561b1e-95e6-4c18-83dd-09930060d986-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418445 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/005399a7-db26-44d0-a0fc-bce24138eb7b-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418455 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/39561b1e-95e6-4c18-83dd-09930060d986-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418464 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf8nx\" (UniqueName: \"kubernetes.io/projected/005399a7-db26-44d0-a0fc-bce24138eb7b-kube-api-access-kf8nx\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.418472 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/005399a7-db26-44d0-a0fc-bce24138eb7b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.519802 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plvrs\" (UniqueName: \"kubernetes.io/projected/6aa3427b-4c78-404f-9cee-2f065504baec-kube-api-access-plvrs\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.519866 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-client-ca\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.519902 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-config\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.519924 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa3427b-4c78-404f-9cee-2f065504baec-serving-cert\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.521019 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-client-ca\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.521065 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-config\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.523991 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa3427b-4c78-404f-9cee-2f065504baec-serving-cert\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.539137 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plvrs\" (UniqueName: \"kubernetes.io/projected/6aa3427b-4c78-404f-9cee-2f065504baec-kube-api-access-plvrs\") pod \"route-controller-manager-669d676fd6-pz6pk\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.557240 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.944754 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk"] Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.992266 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" event={"ID":"39561b1e-95e6-4c18-83dd-09930060d986","Type":"ContainerDied","Data":"7ad71de5e773233b3f73cff196739f67f5e033c493ec6e84a773df37efaf3231"} Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.992599 4692 scope.go:117] "RemoveContainer" containerID="1d8f47335ea3090f1f325d8f1dfd761d242abed8068a348c479b82c0b005b558" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.992314 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.995810 4692 generic.go:334] "Generic (PLEG): container finished" podID="3d4b7447-a6e6-4530-b529-2ec461452791" containerID="cca27779280f6a12e2d3ab6bbcc4fba1f3892575723d534aaea04333cc76bbd1" exitCode=0 Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.995859 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkkhr" event={"ID":"3d4b7447-a6e6-4530-b529-2ec461452791","Type":"ContainerDied","Data":"cca27779280f6a12e2d3ab6bbcc4fba1f3892575723d534aaea04333cc76bbd1"} Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.998741 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" Mar 09 09:22:54 crc kubenswrapper[4692]: I0309 09:22:54.998762 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-84fff57f76-rsxjg" event={"ID":"005399a7-db26-44d0-a0fc-bce24138eb7b","Type":"ContainerDied","Data":"aed99806e2b895cbd1f3c5d7889395bbca314692d09da289bd922a7113bacf1b"} Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.004939 4692 generic.go:334] "Generic (PLEG): container finished" podID="66179a0b-d877-46fb-a854-5426993f4b36" containerID="84aa188df497b4d3620cd2a00ee1bf750251fda196f70faa65fea515f9bcc6a6" exitCode=0 Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.005035 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vbb6" event={"ID":"66179a0b-d877-46fb-a854-5426993f4b36","Type":"ContainerDied","Data":"84aa188df497b4d3620cd2a00ee1bf750251fda196f70faa65fea515f9bcc6a6"} Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.012318 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c61e177f-a1d1-4a8b-a7d7-24035a88f54e","Type":"ContainerStarted","Data":"94af359b51e973959d8c55c9aead27bca67fef420a8629ff6dbe4af9e474c22c"} Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.018594 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" event={"ID":"6aa3427b-4c78-404f-9cee-2f065504baec","Type":"ContainerStarted","Data":"9107f82b5faf3ce99b78f15ea6aba1434c45bf28e06bad0a0c1b17bb4a7f5266"} Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.071382 4692 scope.go:117] "RemoveContainer" containerID="c63066a94b2ada58f76dceecfbc496d3a72fbc538542857d22ba79c17cd4c951" Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.090952 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=4.090928187 podStartE2EDuration="4.090928187s" podCreationTimestamp="2026-03-09 09:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:55.078577889 +0000 UTC m=+175.903313480" watchObservedRunningTime="2026-03-09 09:22:55.090928187 +0000 UTC m=+175.915663768" Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.096228 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-84fff57f76-rsxjg"] Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.106205 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-84fff57f76-rsxjg"] Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.110691 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4"] Mar 09 09:22:55 crc kubenswrapper[4692]: I0309 09:22:55.115498 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f5ccd5b88-b7ps4"] Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.026068 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" event={"ID":"6aa3427b-4c78-404f-9cee-2f065504baec","Type":"ContainerStarted","Data":"508d2ca017b01ca0ed41351995e4839760b1b4f8a9982d81a027ac6d4671e52f"} Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.026256 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.032559 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.049311 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" podStartSLOduration=5.049276897 podStartE2EDuration="5.049276897s" podCreationTimestamp="2026-03-09 09:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:22:56.042145579 +0000 UTC m=+176.866881200" watchObservedRunningTime="2026-03-09 09:22:56.049276897 +0000 UTC m=+176.874012478" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.078931 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="005399a7-db26-44d0-a0fc-bce24138eb7b" path="/var/lib/kubelet/pods/005399a7-db26-44d0-a0fc-bce24138eb7b/volumes" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.079909 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39561b1e-95e6-4c18-83dd-09930060d986" path="/var/lib/kubelet/pods/39561b1e-95e6-4c18-83dd-09930060d986/volumes" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.318209 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-69c5b9bf7f-956hs"] Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.320224 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.323539 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.323797 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.324524 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.324741 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.325054 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.325387 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.329487 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.329799 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69c5b9bf7f-956hs"] Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.457196 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0207afcb-847c-4cdf-b52a-c0a5946a3070-serving-cert\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.457901 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-config\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.458014 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv95g\" (UniqueName: \"kubernetes.io/projected/0207afcb-847c-4cdf-b52a-c0a5946a3070-kube-api-access-zv95g\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.458083 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-proxy-ca-bundles\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.458154 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-client-ca\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.559065 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-config\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.559129 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv95g\" (UniqueName: \"kubernetes.io/projected/0207afcb-847c-4cdf-b52a-c0a5946a3070-kube-api-access-zv95g\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.559170 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-proxy-ca-bundles\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.559250 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-client-ca\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.559287 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0207afcb-847c-4cdf-b52a-c0a5946a3070-serving-cert\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.560264 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-proxy-ca-bundles\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.560336 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-client-ca\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.560523 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-config\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.567256 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0207afcb-847c-4cdf-b52a-c0a5946a3070-serving-cert\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.582070 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv95g\" (UniqueName: \"kubernetes.io/projected/0207afcb-847c-4cdf-b52a-c0a5946a3070-kube-api-access-zv95g\") pod \"controller-manager-69c5b9bf7f-956hs\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:56 crc kubenswrapper[4692]: I0309 09:22:56.651101 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:22:59 crc kubenswrapper[4692]: I0309 09:22:59.283816 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69c5b9bf7f-956hs"] Mar 09 09:23:00 crc kubenswrapper[4692]: I0309 09:23:00.052227 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" event={"ID":"0207afcb-847c-4cdf-b52a-c0a5946a3070","Type":"ContainerStarted","Data":"51281dd25b0e51166e053cee8b5b4b40ff9b5b59fb6f6a07489242eff2192eeb"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.064769 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4bqhb" event={"ID":"73819651-942b-4436-b0b1-7225cfb20ef1","Type":"ContainerStarted","Data":"77afccaf2045943b7772baffcde81fef46c22cefd5b6c5cbd70acc2a3c840789"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.081809 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vbb6" event={"ID":"66179a0b-d877-46fb-a854-5426993f4b36","Type":"ContainerStarted","Data":"0c65c629a52921eac138356df715ea281196bea77620b3ce994f433508226146"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.088045 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerStarted","Data":"56620e11f32ad913c2c09495f85014c2118dd73419dcab9402f83f17cb3b28ce"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.090388 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerStarted","Data":"ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.095433 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" event={"ID":"0207afcb-847c-4cdf-b52a-c0a5946a3070","Type":"ContainerStarted","Data":"011f4877a2f70f9c4dd99f441daf0f94089ccdcce0f4eaffcf98996f1a3d91c0"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.095936 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.098581 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktmgp" event={"ID":"067264d3-8fc2-46cd-8d94-e89f9e25a500","Type":"ContainerStarted","Data":"7bc9bef4716ed588151c98669de90ff155904ec54de77f6e824da71c09e71750"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.100925 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkkhr" event={"ID":"3d4b7447-a6e6-4530-b529-2ec461452791","Type":"ContainerStarted","Data":"cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a"} Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.105579 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.162875 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8vbb6" podStartSLOduration=35.529046443 podStartE2EDuration="47.162856061s" podCreationTimestamp="2026-03-09 09:22:14 +0000 UTC" firstStartedPulling="2026-03-09 09:22:48.877690742 +0000 UTC m=+169.702426323" lastFinishedPulling="2026-03-09 09:23:00.51150036 +0000 UTC m=+181.336235941" observedRunningTime="2026-03-09 09:23:01.158624532 +0000 UTC m=+181.983360133" watchObservedRunningTime="2026-03-09 09:23:01.162856061 +0000 UTC m=+181.987591642" Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.196806 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vkkhr" podStartSLOduration=36.661864266 podStartE2EDuration="48.196787389s" podCreationTimestamp="2026-03-09 09:22:13 +0000 UTC" firstStartedPulling="2026-03-09 09:22:48.854945157 +0000 UTC m=+169.679680738" lastFinishedPulling="2026-03-09 09:23:00.38986828 +0000 UTC m=+181.214603861" observedRunningTime="2026-03-09 09:23:01.195140728 +0000 UTC m=+182.019876309" watchObservedRunningTime="2026-03-09 09:23:01.196787389 +0000 UTC m=+182.021522970" Mar 09 09:23:01 crc kubenswrapper[4692]: I0309 09:23:01.244478 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" podStartSLOduration=10.244456217 podStartE2EDuration="10.244456217s" podCreationTimestamp="2026-03-09 09:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:23:01.221058171 +0000 UTC m=+182.045793772" watchObservedRunningTime="2026-03-09 09:23:01.244456217 +0000 UTC m=+182.069191798" Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.110293 4692 generic.go:334] "Generic (PLEG): container finished" podID="73819651-942b-4436-b0b1-7225cfb20ef1" containerID="77afccaf2045943b7772baffcde81fef46c22cefd5b6c5cbd70acc2a3c840789" exitCode=0 Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.110331 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4bqhb" event={"ID":"73819651-942b-4436-b0b1-7225cfb20ef1","Type":"ContainerDied","Data":"77afccaf2045943b7772baffcde81fef46c22cefd5b6c5cbd70acc2a3c840789"} Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.113834 4692 generic.go:334] "Generic (PLEG): container finished" podID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerID="7bc9bef4716ed588151c98669de90ff155904ec54de77f6e824da71c09e71750" exitCode=0 Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.113995 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktmgp" event={"ID":"067264d3-8fc2-46cd-8d94-e89f9e25a500","Type":"ContainerDied","Data":"7bc9bef4716ed588151c98669de90ff155904ec54de77f6e824da71c09e71750"} Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.116945 4692 generic.go:334] "Generic (PLEG): container finished" podID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerID="56620e11f32ad913c2c09495f85014c2118dd73419dcab9402f83f17cb3b28ce" exitCode=0 Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.117038 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerDied","Data":"56620e11f32ad913c2c09495f85014c2118dd73419dcab9402f83f17cb3b28ce"} Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.120553 4692 generic.go:334] "Generic (PLEG): container finished" podID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerID="ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134" exitCode=0 Mar 09 09:23:02 crc kubenswrapper[4692]: I0309 09:23:02.120658 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerDied","Data":"ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134"} Mar 09 09:23:03 crc kubenswrapper[4692]: I0309 09:23:03.984561 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:23:03 crc kubenswrapper[4692]: I0309 09:23:03.984913 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:23:04 crc kubenswrapper[4692]: I0309 09:23:04.132637 4692 generic.go:334] "Generic (PLEG): container finished" podID="27500a21-baed-4a7c-9af8-e616fee71857" containerID="d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb" exitCode=0 Mar 09 09:23:04 crc kubenswrapper[4692]: I0309 09:23:04.132931 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9nxd" event={"ID":"27500a21-baed-4a7c-9af8-e616fee71857","Type":"ContainerDied","Data":"d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb"} Mar 09 09:23:04 crc kubenswrapper[4692]: I0309 09:23:04.364234 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:23:04 crc kubenswrapper[4692]: I0309 09:23:04.364297 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:23:05 crc kubenswrapper[4692]: I0309 09:23:05.332056 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:23:05 crc kubenswrapper[4692]: I0309 09:23:05.332891 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:23:05 crc kubenswrapper[4692]: I0309 09:23:05.427763 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:23:05 crc kubenswrapper[4692]: I0309 09:23:05.428359 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:23:07 crc kubenswrapper[4692]: I0309 09:23:07.519706 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vbb6"] Mar 09 09:23:07 crc kubenswrapper[4692]: I0309 09:23:07.520324 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8vbb6" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="registry-server" containerID="cri-o://0c65c629a52921eac138356df715ea281196bea77620b3ce994f433508226146" gracePeriod=2 Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.164843 4692 generic.go:334] "Generic (PLEG): container finished" podID="66179a0b-d877-46fb-a854-5426993f4b36" containerID="0c65c629a52921eac138356df715ea281196bea77620b3ce994f433508226146" exitCode=0 Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.164933 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vbb6" event={"ID":"66179a0b-d877-46fb-a854-5426993f4b36","Type":"ContainerDied","Data":"0c65c629a52921eac138356df715ea281196bea77620b3ce994f433508226146"} Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.503337 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.629466 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-utilities\") pod \"66179a0b-d877-46fb-a854-5426993f4b36\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.629614 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhsds\" (UniqueName: \"kubernetes.io/projected/66179a0b-d877-46fb-a854-5426993f4b36-kube-api-access-qhsds\") pod \"66179a0b-d877-46fb-a854-5426993f4b36\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.630382 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-utilities" (OuterVolumeSpecName: "utilities") pod "66179a0b-d877-46fb-a854-5426993f4b36" (UID: "66179a0b-d877-46fb-a854-5426993f4b36"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.630426 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-catalog-content\") pod \"66179a0b-d877-46fb-a854-5426993f4b36\" (UID: \"66179a0b-d877-46fb-a854-5426993f4b36\") " Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.630650 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.643576 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66179a0b-d877-46fb-a854-5426993f4b36-kube-api-access-qhsds" (OuterVolumeSpecName: "kube-api-access-qhsds") pod "66179a0b-d877-46fb-a854-5426993f4b36" (UID: "66179a0b-d877-46fb-a854-5426993f4b36"). InnerVolumeSpecName "kube-api-access-qhsds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.661221 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66179a0b-d877-46fb-a854-5426993f4b36" (UID: "66179a0b-d877-46fb-a854-5426993f4b36"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.731095 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhsds\" (UniqueName: \"kubernetes.io/projected/66179a0b-d877-46fb-a854-5426993f4b36-kube-api-access-qhsds\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:09 crc kubenswrapper[4692]: I0309 09:23:09.731132 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66179a0b-d877-46fb-a854-5426993f4b36-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:10 crc kubenswrapper[4692]: I0309 09:23:10.173009 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qgmsd" event={"ID":"2f27156a-57b6-4230-aa9b-1809cbfb3569","Type":"ContainerStarted","Data":"7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176"} Mar 09 09:23:10 crc kubenswrapper[4692]: I0309 09:23:10.176492 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8vbb6" event={"ID":"66179a0b-d877-46fb-a854-5426993f4b36","Type":"ContainerDied","Data":"4892f652307c25044c45dd41ea567970ecd05cdf431f4b59fd47b39db2128ceb"} Mar 09 09:23:10 crc kubenswrapper[4692]: I0309 09:23:10.176549 4692 scope.go:117] "RemoveContainer" containerID="0c65c629a52921eac138356df715ea281196bea77620b3ce994f433508226146" Mar 09 09:23:10 crc kubenswrapper[4692]: I0309 09:23:10.176704 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8vbb6" Mar 09 09:23:10 crc kubenswrapper[4692]: I0309 09:23:10.194345 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vbb6"] Mar 09 09:23:10 crc kubenswrapper[4692]: I0309 09:23:10.198580 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8vbb6"] Mar 09 09:23:11 crc kubenswrapper[4692]: I0309 09:23:11.182801 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktmgp" event={"ID":"067264d3-8fc2-46cd-8d94-e89f9e25a500","Type":"ContainerStarted","Data":"efea09f529044bf3f09eca867fd99b9e990d01df4cf75b431b94df5261d9fb7b"} Mar 09 09:23:11 crc kubenswrapper[4692]: I0309 09:23:11.184192 4692 generic.go:334] "Generic (PLEG): container finished" podID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerID="7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176" exitCode=0 Mar 09 09:23:11 crc kubenswrapper[4692]: I0309 09:23:11.184226 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qgmsd" event={"ID":"2f27156a-57b6-4230-aa9b-1809cbfb3569","Type":"ContainerDied","Data":"7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176"} Mar 09 09:23:11 crc kubenswrapper[4692]: I0309 09:23:11.382505 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-69c5b9bf7f-956hs"] Mar 09 09:23:11 crc kubenswrapper[4692]: I0309 09:23:11.382767 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" podUID="0207afcb-847c-4cdf-b52a-c0a5946a3070" containerName="controller-manager" containerID="cri-o://011f4877a2f70f9c4dd99f441daf0f94089ccdcce0f4eaffcf98996f1a3d91c0" gracePeriod=30 Mar 09 09:23:11 crc kubenswrapper[4692]: I0309 09:23:11.414907 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk"] Mar 09 09:23:11 crc kubenswrapper[4692]: I0309 09:23:11.415151 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" podUID="6aa3427b-4c78-404f-9cee-2f065504baec" containerName="route-controller-manager" containerID="cri-o://508d2ca017b01ca0ed41351995e4839760b1b4f8a9982d81a027ac6d4671e52f" gracePeriod=30 Mar 09 09:23:12 crc kubenswrapper[4692]: I0309 09:23:12.078786 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66179a0b-d877-46fb-a854-5426993f4b36" path="/var/lib/kubelet/pods/66179a0b-d877-46fb-a854-5426993f4b36/volumes" Mar 09 09:23:12 crc kubenswrapper[4692]: I0309 09:23:12.206368 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ktmgp" podStartSLOduration=6.077935203 podStartE2EDuration="1m1.206352553s" podCreationTimestamp="2026-03-09 09:22:11 +0000 UTC" firstStartedPulling="2026-03-09 09:22:13.49621306 +0000 UTC m=+134.320948641" lastFinishedPulling="2026-03-09 09:23:08.62463041 +0000 UTC m=+189.449365991" observedRunningTime="2026-03-09 09:23:12.203238118 +0000 UTC m=+193.027973699" watchObservedRunningTime="2026-03-09 09:23:12.206352553 +0000 UTC m=+193.031088134" Mar 09 09:23:13 crc kubenswrapper[4692]: I0309 09:23:13.195932 4692 generic.go:334] "Generic (PLEG): container finished" podID="0207afcb-847c-4cdf-b52a-c0a5946a3070" containerID="011f4877a2f70f9c4dd99f441daf0f94089ccdcce0f4eaffcf98996f1a3d91c0" exitCode=0 Mar 09 09:23:13 crc kubenswrapper[4692]: I0309 09:23:13.196027 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" event={"ID":"0207afcb-847c-4cdf-b52a-c0a5946a3070","Type":"ContainerDied","Data":"011f4877a2f70f9c4dd99f441daf0f94089ccdcce0f4eaffcf98996f1a3d91c0"} Mar 09 09:23:13 crc kubenswrapper[4692]: I0309 09:23:13.197897 4692 generic.go:334] "Generic (PLEG): container finished" podID="6aa3427b-4c78-404f-9cee-2f065504baec" containerID="508d2ca017b01ca0ed41351995e4839760b1b4f8a9982d81a027ac6d4671e52f" exitCode=0 Mar 09 09:23:13 crc kubenswrapper[4692]: I0309 09:23:13.197988 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" event={"ID":"6aa3427b-4c78-404f-9cee-2f065504baec","Type":"ContainerDied","Data":"508d2ca017b01ca0ed41351995e4839760b1b4f8a9982d81a027ac6d4671e52f"} Mar 09 09:23:13 crc kubenswrapper[4692]: I0309 09:23:13.866262 4692 scope.go:117] "RemoveContainer" containerID="84aa188df497b4d3620cd2a00ee1bf750251fda196f70faa65fea515f9bcc6a6" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.397460 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.402539 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423218 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-87c746796-wbqnz"] Mar 09 09:23:14 crc kubenswrapper[4692]: E0309 09:23:14.423434 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="extract-content" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423446 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="extract-content" Mar 09 09:23:14 crc kubenswrapper[4692]: E0309 09:23:14.423456 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa3427b-4c78-404f-9cee-2f065504baec" containerName="route-controller-manager" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423463 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa3427b-4c78-404f-9cee-2f065504baec" containerName="route-controller-manager" Mar 09 09:23:14 crc kubenswrapper[4692]: E0309 09:23:14.423473 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="extract-utilities" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423479 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="extract-utilities" Mar 09 09:23:14 crc kubenswrapper[4692]: E0309 09:23:14.423488 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0207afcb-847c-4cdf-b52a-c0a5946a3070" containerName="controller-manager" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423494 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="0207afcb-847c-4cdf-b52a-c0a5946a3070" containerName="controller-manager" Mar 09 09:23:14 crc kubenswrapper[4692]: E0309 09:23:14.423503 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="registry-server" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423509 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="registry-server" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423610 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa3427b-4c78-404f-9cee-2f065504baec" containerName="route-controller-manager" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423629 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="0207afcb-847c-4cdf-b52a-c0a5946a3070" containerName="controller-manager" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.423637 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="66179a0b-d877-46fb-a854-5426993f4b36" containerName="registry-server" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.424410 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.435368 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87c746796-wbqnz"] Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487202 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-config\") pod \"6aa3427b-4c78-404f-9cee-2f065504baec\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487278 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-proxy-ca-bundles\") pod \"0207afcb-847c-4cdf-b52a-c0a5946a3070\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487306 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-client-ca\") pod \"6aa3427b-4c78-404f-9cee-2f065504baec\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487323 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-client-ca\") pod \"0207afcb-847c-4cdf-b52a-c0a5946a3070\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487343 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa3427b-4c78-404f-9cee-2f065504baec-serving-cert\") pod \"6aa3427b-4c78-404f-9cee-2f065504baec\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487375 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0207afcb-847c-4cdf-b52a-c0a5946a3070-serving-cert\") pod \"0207afcb-847c-4cdf-b52a-c0a5946a3070\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487420 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plvrs\" (UniqueName: \"kubernetes.io/projected/6aa3427b-4c78-404f-9cee-2f065504baec-kube-api-access-plvrs\") pod \"6aa3427b-4c78-404f-9cee-2f065504baec\" (UID: \"6aa3427b-4c78-404f-9cee-2f065504baec\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487447 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv95g\" (UniqueName: \"kubernetes.io/projected/0207afcb-847c-4cdf-b52a-c0a5946a3070-kube-api-access-zv95g\") pod \"0207afcb-847c-4cdf-b52a-c0a5946a3070\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487466 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-config\") pod \"0207afcb-847c-4cdf-b52a-c0a5946a3070\" (UID: \"0207afcb-847c-4cdf-b52a-c0a5946a3070\") " Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487626 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-proxy-ca-bundles\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487656 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-client-ca\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487682 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzpxd\" (UniqueName: \"kubernetes.io/projected/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-kube-api-access-pzpxd\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487702 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-serving-cert\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.487750 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-config\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.488898 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-config" (OuterVolumeSpecName: "config") pod "6aa3427b-4c78-404f-9cee-2f065504baec" (UID: "6aa3427b-4c78-404f-9cee-2f065504baec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.488907 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-client-ca" (OuterVolumeSpecName: "client-ca") pod "6aa3427b-4c78-404f-9cee-2f065504baec" (UID: "6aa3427b-4c78-404f-9cee-2f065504baec"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.490889 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-config" (OuterVolumeSpecName: "config") pod "0207afcb-847c-4cdf-b52a-c0a5946a3070" (UID: "0207afcb-847c-4cdf-b52a-c0a5946a3070"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.491480 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0207afcb-847c-4cdf-b52a-c0a5946a3070" (UID: "0207afcb-847c-4cdf-b52a-c0a5946a3070"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.491711 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-client-ca" (OuterVolumeSpecName: "client-ca") pod "0207afcb-847c-4cdf-b52a-c0a5946a3070" (UID: "0207afcb-847c-4cdf-b52a-c0a5946a3070"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.493800 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0207afcb-847c-4cdf-b52a-c0a5946a3070-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0207afcb-847c-4cdf-b52a-c0a5946a3070" (UID: "0207afcb-847c-4cdf-b52a-c0a5946a3070"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.493939 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa3427b-4c78-404f-9cee-2f065504baec-kube-api-access-plvrs" (OuterVolumeSpecName: "kube-api-access-plvrs") pod "6aa3427b-4c78-404f-9cee-2f065504baec" (UID: "6aa3427b-4c78-404f-9cee-2f065504baec"). InnerVolumeSpecName "kube-api-access-plvrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.497270 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0207afcb-847c-4cdf-b52a-c0a5946a3070-kube-api-access-zv95g" (OuterVolumeSpecName: "kube-api-access-zv95g") pod "0207afcb-847c-4cdf-b52a-c0a5946a3070" (UID: "0207afcb-847c-4cdf-b52a-c0a5946a3070"). InnerVolumeSpecName "kube-api-access-zv95g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.500258 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa3427b-4c78-404f-9cee-2f065504baec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6aa3427b-4c78-404f-9cee-2f065504baec" (UID: "6aa3427b-4c78-404f-9cee-2f065504baec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.588808 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-config\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.588866 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-proxy-ca-bundles\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.588897 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-client-ca\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.588924 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzpxd\" (UniqueName: \"kubernetes.io/projected/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-kube-api-access-pzpxd\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.588944 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-serving-cert\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.588999 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zv95g\" (UniqueName: \"kubernetes.io/projected/0207afcb-847c-4cdf-b52a-c0a5946a3070-kube-api-access-zv95g\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589010 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589022 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589032 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589039 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6aa3427b-4c78-404f-9cee-2f065504baec-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589046 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0207afcb-847c-4cdf-b52a-c0a5946a3070-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589055 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa3427b-4c78-404f-9cee-2f065504baec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589063 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0207afcb-847c-4cdf-b52a-c0a5946a3070-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.589071 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plvrs\" (UniqueName: \"kubernetes.io/projected/6aa3427b-4c78-404f-9cee-2f065504baec-kube-api-access-plvrs\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.590532 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-client-ca\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.592179 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-proxy-ca-bundles\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.593288 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-serving-cert\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.593403 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-config\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.603739 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzpxd\" (UniqueName: \"kubernetes.io/projected/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-kube-api-access-pzpxd\") pod \"controller-manager-87c746796-wbqnz\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.731638 4692 scope.go:117] "RemoveContainer" containerID="d02ca25939ba28102a59713650f778d6e75c59d85875075436a4ffd69ee7c0c4" Mar 09 09:23:14 crc kubenswrapper[4692]: I0309 09:23:14.755046 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.211470 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" event={"ID":"0207afcb-847c-4cdf-b52a-c0a5946a3070","Type":"ContainerDied","Data":"51281dd25b0e51166e053cee8b5b4b40ff9b5b59fb6f6a07489242eff2192eeb"} Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.211535 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69c5b9bf7f-956hs" Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.213579 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" event={"ID":"6aa3427b-4c78-404f-9cee-2f065504baec","Type":"ContainerDied","Data":"9107f82b5faf3ce99b78f15ea6aba1434c45bf28e06bad0a0c1b17bb4a7f5266"} Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.213643 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk" Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.242131 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk"] Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.245888 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-669d676fd6-pz6pk"] Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.259869 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-69c5b9bf7f-956hs"] Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.263909 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-69c5b9bf7f-956hs"] Mar 09 09:23:15 crc kubenswrapper[4692]: I0309 09:23:15.915092 4692 scope.go:117] "RemoveContainer" containerID="011f4877a2f70f9c4dd99f441daf0f94089ccdcce0f4eaffcf98996f1a3d91c0" Mar 09 09:23:16 crc kubenswrapper[4692]: I0309 09:23:16.079149 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0207afcb-847c-4cdf-b52a-c0a5946a3070" path="/var/lib/kubelet/pods/0207afcb-847c-4cdf-b52a-c0a5946a3070/volumes" Mar 09 09:23:16 crc kubenswrapper[4692]: I0309 09:23:16.079894 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa3427b-4c78-404f-9cee-2f065504baec" path="/var/lib/kubelet/pods/6aa3427b-4c78-404f-9cee-2f065504baec/volumes" Mar 09 09:23:16 crc kubenswrapper[4692]: I0309 09:23:16.782444 4692 scope.go:117] "RemoveContainer" containerID="508d2ca017b01ca0ed41351995e4839760b1b4f8a9982d81a027ac6d4671e52f" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.333581 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx"] Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.334693 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.339390 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.339984 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.340152 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.340691 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.340824 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.342482 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.343790 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx"] Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.422916 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-client-ca\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.423001 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-config\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.423028 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ad5398b-031a-4fe6-a264-9e0df09ade1a-serving-cert\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.423078 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7t6z\" (UniqueName: \"kubernetes.io/projected/4ad5398b-031a-4fe6-a264-9e0df09ade1a-kube-api-access-f7t6z\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.524456 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-client-ca\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.524562 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-config\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.524595 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ad5398b-031a-4fe6-a264-9e0df09ade1a-serving-cert\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.524644 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7t6z\" (UniqueName: \"kubernetes.io/projected/4ad5398b-031a-4fe6-a264-9e0df09ade1a-kube-api-access-f7t6z\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.525504 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-client-ca\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.525783 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-config\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.528580 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ad5398b-031a-4fe6-a264-9e0df09ade1a-serving-cert\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.540676 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7t6z\" (UniqueName: \"kubernetes.io/projected/4ad5398b-031a-4fe6-a264-9e0df09ade1a-kube-api-access-f7t6z\") pod \"route-controller-manager-665bb7464f-hq9xx\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:17 crc kubenswrapper[4692]: I0309 09:23:17.653257 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.010479 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-87c746796-wbqnz"] Mar 09 09:23:18 crc kubenswrapper[4692]: W0309 09:23:18.019653 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d0ec70a_c384_4fe0_ae86_3e0da3f3637e.slice/crio-3498e75169645cc13a511b603124cdd5e2755c25f34ef9b6a6395f55472a6920 WatchSource:0}: Error finding container 3498e75169645cc13a511b603124cdd5e2755c25f34ef9b6a6395f55472a6920: Status 404 returned error can't find the container with id 3498e75169645cc13a511b603124cdd5e2755c25f34ef9b6a6395f55472a6920 Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.109488 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx"] Mar 09 09:23:18 crc kubenswrapper[4692]: W0309 09:23:18.120322 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad5398b_031a_4fe6_a264_9e0df09ade1a.slice/crio-b277b4eb10c38603397ed9aadad292452d2c51cfd0c9824256c05100c9b3a333 WatchSource:0}: Error finding container b277b4eb10c38603397ed9aadad292452d2c51cfd0c9824256c05100c9b3a333: Status 404 returned error can't find the container with id b277b4eb10c38603397ed9aadad292452d2c51cfd0c9824256c05100c9b3a333 Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.237762 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4bqhb" event={"ID":"73819651-942b-4436-b0b1-7225cfb20ef1","Type":"ContainerStarted","Data":"6ec1ecbb6a65d8760942f519a7440b6a4dcfb4e6184b9b199bc82504b9635683"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.244427 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qgmsd" event={"ID":"2f27156a-57b6-4230-aa9b-1809cbfb3569","Type":"ContainerStarted","Data":"de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.245528 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" event={"ID":"4ad5398b-031a-4fe6-a264-9e0df09ade1a","Type":"ContainerStarted","Data":"b277b4eb10c38603397ed9aadad292452d2c51cfd0c9824256c05100c9b3a333"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.249995 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9nxd" event={"ID":"27500a21-baed-4a7c-9af8-e616fee71857","Type":"ContainerStarted","Data":"fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.252283 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerStarted","Data":"2f11fd8acf0cbe68dc0911c7764a229e4c763847bce01b88fa085f3dd4ae6726"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.254644 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerStarted","Data":"e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.256736 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" event={"ID":"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e","Type":"ContainerStarted","Data":"8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.256792 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" event={"ID":"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e","Type":"ContainerStarted","Data":"3498e75169645cc13a511b603124cdd5e2755c25f34ef9b6a6395f55472a6920"} Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.257012 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.259541 4692 patch_prober.go:28] interesting pod/controller-manager-87c746796-wbqnz container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" start-of-body= Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.259607 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.270706 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4bqhb" podStartSLOduration=4.191909118 podStartE2EDuration="1m7.27068848s" podCreationTimestamp="2026-03-09 09:22:11 +0000 UTC" firstStartedPulling="2026-03-09 09:22:13.51745079 +0000 UTC m=+134.342186371" lastFinishedPulling="2026-03-09 09:23:16.596230152 +0000 UTC m=+197.420965733" observedRunningTime="2026-03-09 09:23:18.268979228 +0000 UTC m=+199.093714829" watchObservedRunningTime="2026-03-09 09:23:18.27068848 +0000 UTC m=+199.095424071" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.293919 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qgmsd" podStartSLOduration=3.245873906 podStartE2EDuration="1m7.293895941s" podCreationTimestamp="2026-03-09 09:22:11 +0000 UTC" firstStartedPulling="2026-03-09 09:22:13.531591422 +0000 UTC m=+134.356327003" lastFinishedPulling="2026-03-09 09:23:17.579613467 +0000 UTC m=+198.404349038" observedRunningTime="2026-03-09 09:23:18.290459515 +0000 UTC m=+199.115195116" watchObservedRunningTime="2026-03-09 09:23:18.293895941 +0000 UTC m=+199.118631522" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.314939 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q9nxd" podStartSLOduration=2.516843657 podStartE2EDuration="1m6.314921095s" podCreationTimestamp="2026-03-09 09:22:12 +0000 UTC" firstStartedPulling="2026-03-09 09:22:13.498935714 +0000 UTC m=+134.323671295" lastFinishedPulling="2026-03-09 09:23:17.297013152 +0000 UTC m=+198.121748733" observedRunningTime="2026-03-09 09:23:18.312314865 +0000 UTC m=+199.137050456" watchObservedRunningTime="2026-03-09 09:23:18.314921095 +0000 UTC m=+199.139656676" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.342609 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-998gk" podStartSLOduration=36.852609096 podStartE2EDuration="1m3.342590262s" podCreationTimestamp="2026-03-09 09:22:15 +0000 UTC" firstStartedPulling="2026-03-09 09:22:48.858618879 +0000 UTC m=+169.683354460" lastFinishedPulling="2026-03-09 09:23:15.348600045 +0000 UTC m=+196.173335626" observedRunningTime="2026-03-09 09:23:18.341028944 +0000 UTC m=+199.165764545" watchObservedRunningTime="2026-03-09 09:23:18.342590262 +0000 UTC m=+199.167325843" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.368740 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pnf7m" podStartSLOduration=38.521542995 podStartE2EDuration="1m4.368704221s" podCreationTimestamp="2026-03-09 09:22:14 +0000 UTC" firstStartedPulling="2026-03-09 09:22:48.884350336 +0000 UTC m=+169.709085917" lastFinishedPulling="2026-03-09 09:23:14.731511562 +0000 UTC m=+195.556247143" observedRunningTime="2026-03-09 09:23:18.364979497 +0000 UTC m=+199.189715098" watchObservedRunningTime="2026-03-09 09:23:18.368704221 +0000 UTC m=+199.193439822" Mar 09 09:23:18 crc kubenswrapper[4692]: I0309 09:23:18.387459 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" podStartSLOduration=7.3874380649999996 podStartE2EDuration="7.387438065s" podCreationTimestamp="2026-03-09 09:23:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:23:18.385477905 +0000 UTC m=+199.210213486" watchObservedRunningTime="2026-03-09 09:23:18.387438065 +0000 UTC m=+199.212173666" Mar 09 09:23:19 crc kubenswrapper[4692]: I0309 09:23:19.263464 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" event={"ID":"4ad5398b-031a-4fe6-a264-9e0df09ade1a","Type":"ContainerStarted","Data":"83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7"} Mar 09 09:23:19 crc kubenswrapper[4692]: I0309 09:23:19.271434 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:19 crc kubenswrapper[4692]: I0309 09:23:19.285296 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" podStartSLOduration=8.285265279 podStartE2EDuration="8.285265279s" podCreationTimestamp="2026-03-09 09:23:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:23:19.283815664 +0000 UTC m=+200.108551245" watchObservedRunningTime="2026-03-09 09:23:19.285265279 +0000 UTC m=+200.110000880" Mar 09 09:23:20 crc kubenswrapper[4692]: I0309 09:23:20.267821 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:20 crc kubenswrapper[4692]: I0309 09:23:20.275264 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:20 crc kubenswrapper[4692]: I0309 09:23:20.762756 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-pbl6c"] Mar 09 09:23:21 crc kubenswrapper[4692]: I0309 09:23:21.779358 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:23:21 crc kubenswrapper[4692]: I0309 09:23:21.779410 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:23:21 crc kubenswrapper[4692]: I0309 09:23:21.818593 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:23:21 crc kubenswrapper[4692]: I0309 09:23:21.981775 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:23:21 crc kubenswrapper[4692]: I0309 09:23:21.982139 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.035932 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.224547 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.225352 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.262802 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.330024 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.332206 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.332749 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.380438 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.380476 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:23:22 crc kubenswrapper[4692]: I0309 09:23:22.415687 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:23:23 crc kubenswrapper[4692]: I0309 09:23:23.325098 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qgmsd"] Mar 09 09:23:23 crc kubenswrapper[4692]: I0309 09:23:23.329986 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.288897 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qgmsd" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="registry-server" containerID="cri-o://de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb" gracePeriod=2 Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.322075 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q9nxd"] Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.723668 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.831134 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgtzd\" (UniqueName: \"kubernetes.io/projected/2f27156a-57b6-4230-aa9b-1809cbfb3569-kube-api-access-jgtzd\") pod \"2f27156a-57b6-4230-aa9b-1809cbfb3569\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.831201 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-utilities\") pod \"2f27156a-57b6-4230-aa9b-1809cbfb3569\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.831380 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-catalog-content\") pod \"2f27156a-57b6-4230-aa9b-1809cbfb3569\" (UID: \"2f27156a-57b6-4230-aa9b-1809cbfb3569\") " Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.832940 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-utilities" (OuterVolumeSpecName: "utilities") pod "2f27156a-57b6-4230-aa9b-1809cbfb3569" (UID: "2f27156a-57b6-4230-aa9b-1809cbfb3569"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.836325 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f27156a-57b6-4230-aa9b-1809cbfb3569-kube-api-access-jgtzd" (OuterVolumeSpecName: "kube-api-access-jgtzd") pod "2f27156a-57b6-4230-aa9b-1809cbfb3569" (UID: "2f27156a-57b6-4230-aa9b-1809cbfb3569"). InnerVolumeSpecName "kube-api-access-jgtzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.895628 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f27156a-57b6-4230-aa9b-1809cbfb3569" (UID: "2f27156a-57b6-4230-aa9b-1809cbfb3569"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.932442 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.932681 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgtzd\" (UniqueName: \"kubernetes.io/projected/2f27156a-57b6-4230-aa9b-1809cbfb3569-kube-api-access-jgtzd\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.932752 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f27156a-57b6-4230-aa9b-1809cbfb3569-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.985498 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:23:24 crc kubenswrapper[4692]: I0309 09:23:24.985560 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.021794 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.295654 4692 generic.go:334] "Generic (PLEG): container finished" podID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerID="de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb" exitCode=0 Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.295770 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qgmsd" event={"ID":"2f27156a-57b6-4230-aa9b-1809cbfb3569","Type":"ContainerDied","Data":"de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb"} Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.295838 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qgmsd" event={"ID":"2f27156a-57b6-4230-aa9b-1809cbfb3569","Type":"ContainerDied","Data":"2fd137a190bddd010c8c6e2c625b368d2503c19f62836ace20067310dd96db5b"} Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.295861 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q9nxd" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="registry-server" containerID="cri-o://fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c" gracePeriod=2 Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.295864 4692 scope.go:117] "RemoveContainer" containerID="de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.296209 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qgmsd" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.321970 4692 scope.go:117] "RemoveContainer" containerID="7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.343296 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qgmsd"] Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.346574 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.349706 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qgmsd"] Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.354436 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.354635 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.351343 4692 scope.go:117] "RemoveContainer" containerID="411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.394055 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.411502 4692 scope.go:117] "RemoveContainer" containerID="de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb" Mar 09 09:23:25 crc kubenswrapper[4692]: E0309 09:23:25.412065 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb\": container with ID starting with de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb not found: ID does not exist" containerID="de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.412098 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb"} err="failed to get container status \"de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb\": rpc error: code = NotFound desc = could not find container \"de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb\": container with ID starting with de809cd641516a18925250ad8c1ea511fa9ac63f1f443dfe24363ae877d873fb not found: ID does not exist" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.412122 4692 scope.go:117] "RemoveContainer" containerID="7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176" Mar 09 09:23:25 crc kubenswrapper[4692]: E0309 09:23:25.412513 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176\": container with ID starting with 7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176 not found: ID does not exist" containerID="7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.412537 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176"} err="failed to get container status \"7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176\": rpc error: code = NotFound desc = could not find container \"7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176\": container with ID starting with 7b3a0dec68a7948bafc5d66542d3f13056476de4c894111c12981c4c0ecd1176 not found: ID does not exist" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.412554 4692 scope.go:117] "RemoveContainer" containerID="411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d" Mar 09 09:23:25 crc kubenswrapper[4692]: E0309 09:23:25.414571 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d\": container with ID starting with 411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d not found: ID does not exist" containerID="411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.414647 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d"} err="failed to get container status \"411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d\": rpc error: code = NotFound desc = could not find container \"411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d\": container with ID starting with 411464d1c4672cd43654023e3021ee7275909d434714314693b123914f28fe3d not found: ID does not exist" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.769448 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.947981 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9w5h\" (UniqueName: \"kubernetes.io/projected/27500a21-baed-4a7c-9af8-e616fee71857-kube-api-access-w9w5h\") pod \"27500a21-baed-4a7c-9af8-e616fee71857\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.948045 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-catalog-content\") pod \"27500a21-baed-4a7c-9af8-e616fee71857\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.948105 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-utilities\") pod \"27500a21-baed-4a7c-9af8-e616fee71857\" (UID: \"27500a21-baed-4a7c-9af8-e616fee71857\") " Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.949233 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-utilities" (OuterVolumeSpecName: "utilities") pod "27500a21-baed-4a7c-9af8-e616fee71857" (UID: "27500a21-baed-4a7c-9af8-e616fee71857"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.952501 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27500a21-baed-4a7c-9af8-e616fee71857-kube-api-access-w9w5h" (OuterVolumeSpecName: "kube-api-access-w9w5h") pod "27500a21-baed-4a7c-9af8-e616fee71857" (UID: "27500a21-baed-4a7c-9af8-e616fee71857"). InnerVolumeSpecName "kube-api-access-w9w5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:25 crc kubenswrapper[4692]: I0309 09:23:25.993394 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27500a21-baed-4a7c-9af8-e616fee71857" (UID: "27500a21-baed-4a7c-9af8-e616fee71857"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.049495 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9w5h\" (UniqueName: \"kubernetes.io/projected/27500a21-baed-4a7c-9af8-e616fee71857-kube-api-access-w9w5h\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.049540 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.049552 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27500a21-baed-4a7c-9af8-e616fee71857-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.079518 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" path="/var/lib/kubelet/pods/2f27156a-57b6-4230-aa9b-1809cbfb3569/volumes" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.279744 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.309771 4692 generic.go:334] "Generic (PLEG): container finished" podID="27500a21-baed-4a7c-9af8-e616fee71857" containerID="fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c" exitCode=0 Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.310332 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9nxd" event={"ID":"27500a21-baed-4a7c-9af8-e616fee71857","Type":"ContainerDied","Data":"fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c"} Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.310355 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q9nxd" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.310399 4692 scope.go:117] "RemoveContainer" containerID="fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.310386 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q9nxd" event={"ID":"27500a21-baed-4a7c-9af8-e616fee71857","Type":"ContainerDied","Data":"c2dd6eb569055c8b33157bfdf8bb6206a92145ba93cf980592e31347c6c2f70a"} Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.331283 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q9nxd"] Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.333337 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q9nxd"] Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.337006 4692 scope.go:117] "RemoveContainer" containerID="d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.356799 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.359649 4692 scope.go:117] "RemoveContainer" containerID="75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.376417 4692 scope.go:117] "RemoveContainer" containerID="fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c" Mar 09 09:23:26 crc kubenswrapper[4692]: E0309 09:23:26.377467 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c\": container with ID starting with fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c not found: ID does not exist" containerID="fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.377515 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c"} err="failed to get container status \"fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c\": rpc error: code = NotFound desc = could not find container \"fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c\": container with ID starting with fb01929a74dae6c659545dc6d8853cda1bf65aa80f5ed0dd83f1e735b5cd6b5c not found: ID does not exist" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.377542 4692 scope.go:117] "RemoveContainer" containerID="d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb" Mar 09 09:23:26 crc kubenswrapper[4692]: E0309 09:23:26.377896 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb\": container with ID starting with d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb not found: ID does not exist" containerID="d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.377935 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb"} err="failed to get container status \"d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb\": rpc error: code = NotFound desc = could not find container \"d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb\": container with ID starting with d861746ddc943425004e235efdf0ad7824ce43d67e381ce475c18f89465a76bb not found: ID does not exist" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.377967 4692 scope.go:117] "RemoveContainer" containerID="75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c" Mar 09 09:23:26 crc kubenswrapper[4692]: E0309 09:23:26.378893 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c\": container with ID starting with 75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c not found: ID does not exist" containerID="75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c" Mar 09 09:23:26 crc kubenswrapper[4692]: I0309 09:23:26.378921 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c"} err="failed to get container status \"75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c\": rpc error: code = NotFound desc = could not find container \"75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c\": container with ID starting with 75c088ab12007fab3f21393f6d26272ee290483b52937e48beb82e3355e5d56c not found: ID does not exist" Mar 09 09:23:27 crc kubenswrapper[4692]: I0309 09:23:27.723134 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-998gk"] Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.078747 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27500a21-baed-4a7c-9af8-e616fee71857" path="/var/lib/kubelet/pods/27500a21-baed-4a7c-9af8-e616fee71857/volumes" Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.323377 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-998gk" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="registry-server" containerID="cri-o://e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528" gracePeriod=2 Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.825000 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.991158 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-utilities\") pod \"fd840400-6e1f-4556-b60b-0cbacff31f1d\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.991238 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-catalog-content\") pod \"fd840400-6e1f-4556-b60b-0cbacff31f1d\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.991300 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t5jn\" (UniqueName: \"kubernetes.io/projected/fd840400-6e1f-4556-b60b-0cbacff31f1d-kube-api-access-4t5jn\") pod \"fd840400-6e1f-4556-b60b-0cbacff31f1d\" (UID: \"fd840400-6e1f-4556-b60b-0cbacff31f1d\") " Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.992138 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-utilities" (OuterVolumeSpecName: "utilities") pod "fd840400-6e1f-4556-b60b-0cbacff31f1d" (UID: "fd840400-6e1f-4556-b60b-0cbacff31f1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:28 crc kubenswrapper[4692]: I0309 09:23:28.996084 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd840400-6e1f-4556-b60b-0cbacff31f1d-kube-api-access-4t5jn" (OuterVolumeSpecName: "kube-api-access-4t5jn") pod "fd840400-6e1f-4556-b60b-0cbacff31f1d" (UID: "fd840400-6e1f-4556-b60b-0cbacff31f1d"). InnerVolumeSpecName "kube-api-access-4t5jn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.092604 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.092891 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t5jn\" (UniqueName: \"kubernetes.io/projected/fd840400-6e1f-4556-b60b-0cbacff31f1d-kube-api-access-4t5jn\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.137250 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd840400-6e1f-4556-b60b-0cbacff31f1d" (UID: "fd840400-6e1f-4556-b60b-0cbacff31f1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.193751 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd840400-6e1f-4556-b60b-0cbacff31f1d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.331355 4692 generic.go:334] "Generic (PLEG): container finished" podID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerID="e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528" exitCode=0 Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.331402 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerDied","Data":"e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528"} Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.331427 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-998gk" event={"ID":"fd840400-6e1f-4556-b60b-0cbacff31f1d","Type":"ContainerDied","Data":"d5a5daa50f56a4026015ac63b025dab64e3579bf76c4b395d5a3c67048445090"} Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.331444 4692 scope.go:117] "RemoveContainer" containerID="e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.331610 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-998gk" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.349538 4692 scope.go:117] "RemoveContainer" containerID="ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.358248 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-998gk"] Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.363608 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-998gk"] Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.371710 4692 scope.go:117] "RemoveContainer" containerID="4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.386070 4692 scope.go:117] "RemoveContainer" containerID="e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528" Mar 09 09:23:29 crc kubenswrapper[4692]: E0309 09:23:29.386510 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528\": container with ID starting with e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528 not found: ID does not exist" containerID="e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.386538 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528"} err="failed to get container status \"e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528\": rpc error: code = NotFound desc = could not find container \"e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528\": container with ID starting with e44740b595842ed436ad96783ccaf55ac9f8c7152881f8570c94573db4f94528 not found: ID does not exist" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.386561 4692 scope.go:117] "RemoveContainer" containerID="ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134" Mar 09 09:23:29 crc kubenswrapper[4692]: E0309 09:23:29.386823 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134\": container with ID starting with ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134 not found: ID does not exist" containerID="ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.386848 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134"} err="failed to get container status \"ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134\": rpc error: code = NotFound desc = could not find container \"ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134\": container with ID starting with ef7211d9f1ad17822657621c9f497a1d5d7743b29868242805e659e1d20d4134 not found: ID does not exist" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.386861 4692 scope.go:117] "RemoveContainer" containerID="4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba" Mar 09 09:23:29 crc kubenswrapper[4692]: E0309 09:23:29.387143 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba\": container with ID starting with 4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba not found: ID does not exist" containerID="4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba" Mar 09 09:23:29 crc kubenswrapper[4692]: I0309 09:23:29.387201 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba"} err="failed to get container status \"4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba\": rpc error: code = NotFound desc = could not find container \"4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba\": container with ID starting with 4886db04556f82434c023a2b0eab119ea00d1ebf56d842d604e1a929038422ba not found: ID does not exist" Mar 09 09:23:30 crc kubenswrapper[4692]: I0309 09:23:30.081264 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" path="/var/lib/kubelet/pods/fd840400-6e1f-4556-b60b-0cbacff31f1d/volumes" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.340354 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-87c746796-wbqnz"] Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.340558 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" containerName="controller-manager" containerID="cri-o://8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7" gracePeriod=30 Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.375536 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx"] Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.375794 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" containerName="route-controller-manager" containerID="cri-o://83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7" gracePeriod=30 Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.508241 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad5398b_031a_4fe6_a264_9e0df09ade1a.slice/crio-conmon-83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7.scope\": RecentStats: unable to find data in memory cache]" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.852953 4692 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853464 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="extract-utilities" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853480 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="extract-utilities" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853492 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853499 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853508 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="extract-content" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853516 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="extract-content" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853525 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="extract-utilities" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853531 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="extract-utilities" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853540 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853546 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853557 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="extract-content" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853562 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="extract-content" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853571 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="extract-content" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853595 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="extract-content" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853605 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="extract-utilities" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853613 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="extract-utilities" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.853620 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853626 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853716 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f27156a-57b6-4230-aa9b-1809cbfb3569" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853733 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd840400-6e1f-4556-b60b-0cbacff31f1d" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.853743 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="27500a21-baed-4a7c-9af8-e616fee71857" containerName="registry-server" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.854110 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.855764 4692 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.856025 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967" gracePeriod=15 Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.856037 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245" gracePeriod=15 Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.856137 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370" gracePeriod=15 Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.856183 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234" gracePeriod=15 Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.856228 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa" gracePeriod=15 Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857055 4692 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857515 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857532 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857541 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857549 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857563 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857571 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857583 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857591 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857600 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857608 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857617 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857623 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857634 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857640 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857657 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857663 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857770 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857784 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857792 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857800 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857811 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857820 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857830 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.857941 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.857955 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.858071 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.858197 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.858207 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.858333 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.908905 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.928361 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-client-ca\") pod \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.928438 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ad5398b-031a-4fe6-a264-9e0df09ade1a-serving-cert\") pod \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.928653 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-config\") pod \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.928688 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7t6z\" (UniqueName: \"kubernetes.io/projected/4ad5398b-031a-4fe6-a264-9e0df09ade1a-kube-api-access-f7t6z\") pod \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\" (UID: \"4ad5398b-031a-4fe6-a264-9e0df09ade1a\") " Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929024 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929071 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929103 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929129 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929190 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929211 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929250 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.929298 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.930603 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-config" (OuterVolumeSpecName: "config") pod "4ad5398b-031a-4fe6-a264-9e0df09ade1a" (UID: "4ad5398b-031a-4fe6-a264-9e0df09ade1a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.930802 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-client-ca" (OuterVolumeSpecName: "client-ca") pod "4ad5398b-031a-4fe6-a264-9e0df09ade1a" (UID: "4ad5398b-031a-4fe6-a264-9e0df09ade1a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.937486 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad5398b-031a-4fe6-a264-9e0df09ade1a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4ad5398b-031a-4fe6-a264-9e0df09ade1a" (UID: "4ad5398b-031a-4fe6-a264-9e0df09ade1a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.947391 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad5398b-031a-4fe6-a264-9e0df09ade1a-kube-api-access-f7t6z" (OuterVolumeSpecName: "kube-api-access-f7t6z") pod "4ad5398b-031a-4fe6-a264-9e0df09ade1a" (UID: "4ad5398b-031a-4fe6-a264-9e0df09ade1a"). InnerVolumeSpecName "kube-api-access-f7t6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:31 crc kubenswrapper[4692]: E0309 09:23:31.968138 4692 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:31 crc kubenswrapper[4692]: I0309 09:23:31.976857 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.030634 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-serving-cert\") pod \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031209 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-config\") pod \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031295 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzpxd\" (UniqueName: \"kubernetes.io/projected/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-kube-api-access-pzpxd\") pod \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031319 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-client-ca\") pod \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031432 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-proxy-ca-bundles\") pod \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\" (UID: \"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e\") " Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031582 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031639 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031707 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031738 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031757 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031770 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031793 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031808 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031850 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031863 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7t6z\" (UniqueName: \"kubernetes.io/projected/4ad5398b-031a-4fe6-a264-9e0df09ade1a-kube-api-access-f7t6z\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031874 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad5398b-031a-4fe6-a264-9e0df09ade1a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031883 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ad5398b-031a-4fe6-a264-9e0df09ade1a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.031927 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032367 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-client-ca" (OuterVolumeSpecName: "client-ca") pod "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" (UID: "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032466 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" (UID: "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032480 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032510 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032533 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032521 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-config" (OuterVolumeSpecName: "config") pod "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" (UID: "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032564 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032573 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032588 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.032616 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.035608 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" (UID: "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.035677 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-kube-api-access-pzpxd" (OuterVolumeSpecName: "kube-api-access-pzpxd") pod "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" (UID: "2d0ec70a-c384-4fe0-ae86-3e0da3f3637e"). InnerVolumeSpecName "kube-api-access-pzpxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.133024 4692 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.133064 4692 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.133074 4692 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.133086 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzpxd\" (UniqueName: \"kubernetes.io/projected/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-kube-api-access-pzpxd\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.133095 4692 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.269288 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:32 crc kubenswrapper[4692]: E0309 09:23:32.286847 4692 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.181:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189b21f47ef51a9f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:23:32.286151327 +0000 UTC m=+213.110886908,LastTimestamp:2026-03-09 09:23:32.286151327 +0000 UTC m=+213.110886908,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.347994 4692 generic.go:334] "Generic (PLEG): container finished" podID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" containerID="83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7" exitCode=0 Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.348026 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.348083 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" event={"ID":"4ad5398b-031a-4fe6-a264-9e0df09ade1a","Type":"ContainerDied","Data":"83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7"} Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.348121 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" event={"ID":"4ad5398b-031a-4fe6-a264-9e0df09ade1a","Type":"ContainerDied","Data":"b277b4eb10c38603397ed9aadad292452d2c51cfd0c9824256c05100c9b3a333"} Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.348137 4692 scope.go:117] "RemoveContainer" containerID="83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.350046 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.351389 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.353656 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245" exitCode=0 Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.353691 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370" exitCode=0 Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.353700 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa" exitCode=0 Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.353708 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234" exitCode=2 Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.355127 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b0ea1c8348966fc055407482645ce62045bb605014fb5df84af83736df3a2af4"} Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.356888 4692 generic.go:334] "Generic (PLEG): container finished" podID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" containerID="94af359b51e973959d8c55c9aead27bca67fef420a8629ff6dbe4af9e474c22c" exitCode=0 Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.356927 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c61e177f-a1d1-4a8b-a7d7-24035a88f54e","Type":"ContainerDied","Data":"94af359b51e973959d8c55c9aead27bca67fef420a8629ff6dbe4af9e474c22c"} Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.360077 4692 generic.go:334] "Generic (PLEG): container finished" podID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" containerID="8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7" exitCode=0 Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.360104 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" event={"ID":"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e","Type":"ContainerDied","Data":"8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7"} Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.360121 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" event={"ID":"2d0ec70a-c384-4fe0-ae86-3e0da3f3637e","Type":"ContainerDied","Data":"3498e75169645cc13a511b603124cdd5e2755c25f34ef9b6a6395f55472a6920"} Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.360155 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.415671 4692 scope.go:117] "RemoveContainer" containerID="83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7" Mar 09 09:23:32 crc kubenswrapper[4692]: E0309 09:23:32.416200 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7\": container with ID starting with 83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7 not found: ID does not exist" containerID="83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.416238 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7"} err="failed to get container status \"83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7\": rpc error: code = NotFound desc = could not find container \"83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7\": container with ID starting with 83ebe10c927bdf2abf08fb77f5cd0ddd89c582ba783203872b8905dba2989db7 not found: ID does not exist" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.416273 4692 scope.go:117] "RemoveContainer" containerID="5b5df248f43515be7e53b9d98e253524dde706cab6e475d59e5998fddffbb92f" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.450755 4692 scope.go:117] "RemoveContainer" containerID="8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.484108 4692 scope.go:117] "RemoveContainer" containerID="8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7" Mar 09 09:23:32 crc kubenswrapper[4692]: E0309 09:23:32.484609 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7\": container with ID starting with 8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7 not found: ID does not exist" containerID="8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7" Mar 09 09:23:32 crc kubenswrapper[4692]: I0309 09:23:32.484665 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7"} err="failed to get container status \"8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7\": rpc error: code = NotFound desc = could not find container \"8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7\": container with ID starting with 8e0265b22cbfca67dd2ce1313e47df982e9cc88281617042aa80158943f3fcf7 not found: ID does not exist" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.368812 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"9f2e4096577478b98ac14cd87c69a2d5b8e851ac6682e739bae9f5c7446275a2"} Mar 09 09:23:33 crc kubenswrapper[4692]: E0309 09:23:33.369417 4692 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.372991 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.679307 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.784773 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kubelet-dir\") pod \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.784861 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c61e177f-a1d1-4a8b-a7d7-24035a88f54e" (UID: "c61e177f-a1d1-4a8b-a7d7-24035a88f54e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.784868 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kube-api-access\") pod \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.785006 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-var-lock\") pod \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\" (UID: \"c61e177f-a1d1-4a8b-a7d7-24035a88f54e\") " Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.785149 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-var-lock" (OuterVolumeSpecName: "var-lock") pod "c61e177f-a1d1-4a8b-a7d7-24035a88f54e" (UID: "c61e177f-a1d1-4a8b-a7d7-24035a88f54e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.785861 4692 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-var-lock\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.786115 4692 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.790085 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c61e177f-a1d1-4a8b-a7d7-24035a88f54e" (UID: "c61e177f-a1d1-4a8b-a7d7-24035a88f54e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:33 crc kubenswrapper[4692]: I0309 09:23:33.887300 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c61e177f-a1d1-4a8b-a7d7-24035a88f54e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.233179 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.234228 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.394807 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.394911 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395308 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395355 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395446 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395606 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395740 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395890 4692 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395923 4692 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.395949 4692 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.396684 4692 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967" exitCode=0 Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.396816 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.396824 4692 scope.go:117] "RemoveContainer" containerID="1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.398978 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c61e177f-a1d1-4a8b-a7d7-24035a88f54e","Type":"ContainerDied","Data":"6304cbe8a68ce7c136b8c307f75c2e7a791cd2c35d9a065a35fc41ade603a186"} Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.399013 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6304cbe8a68ce7c136b8c307f75c2e7a791cd2c35d9a065a35fc41ade603a186" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.399047 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 09 09:23:34 crc kubenswrapper[4692]: E0309 09:23:34.399863 4692 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.411942 4692 scope.go:117] "RemoveContainer" containerID="2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.424658 4692 scope.go:117] "RemoveContainer" containerID="aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.437812 4692 scope.go:117] "RemoveContainer" containerID="16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.450262 4692 scope.go:117] "RemoveContainer" containerID="0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.462788 4692 scope.go:117] "RemoveContainer" containerID="19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.478084 4692 scope.go:117] "RemoveContainer" containerID="1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245" Mar 09 09:23:34 crc kubenswrapper[4692]: E0309 09:23:34.478514 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245\": container with ID starting with 1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245 not found: ID does not exist" containerID="1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.478558 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245"} err="failed to get container status \"1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245\": rpc error: code = NotFound desc = could not find container \"1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245\": container with ID starting with 1c5c3edd94270b2266385ef22f4e5115f11dc729419a4ffa51b9be9247391245 not found: ID does not exist" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.478594 4692 scope.go:117] "RemoveContainer" containerID="2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370" Mar 09 09:23:34 crc kubenswrapper[4692]: E0309 09:23:34.478869 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370\": container with ID starting with 2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370 not found: ID does not exist" containerID="2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.478902 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370"} err="failed to get container status \"2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370\": rpc error: code = NotFound desc = could not find container \"2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370\": container with ID starting with 2b0d97157aed1c34e18e8cc3043f6c59bab1575c9d2c031569fdaf108f95d370 not found: ID does not exist" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.478929 4692 scope.go:117] "RemoveContainer" containerID="aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa" Mar 09 09:23:34 crc kubenswrapper[4692]: E0309 09:23:34.479228 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa\": container with ID starting with aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa not found: ID does not exist" containerID="aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.479269 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa"} err="failed to get container status \"aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa\": rpc error: code = NotFound desc = could not find container \"aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa\": container with ID starting with aa9733b0bf5db48b38d8fe328de4933a1b8b1259be0bd213c05afdc3e84fa6fa not found: ID does not exist" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.479294 4692 scope.go:117] "RemoveContainer" containerID="16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234" Mar 09 09:23:34 crc kubenswrapper[4692]: E0309 09:23:34.479554 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234\": container with ID starting with 16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234 not found: ID does not exist" containerID="16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.479587 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234"} err="failed to get container status \"16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234\": rpc error: code = NotFound desc = could not find container \"16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234\": container with ID starting with 16a7d1128f8cafae03225e3dd14acd17ecf38db2102d3c4b9884bdff87b10234 not found: ID does not exist" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.479610 4692 scope.go:117] "RemoveContainer" containerID="0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967" Mar 09 09:23:34 crc kubenswrapper[4692]: E0309 09:23:34.479842 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967\": container with ID starting with 0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967 not found: ID does not exist" containerID="0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.479879 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967"} err="failed to get container status \"0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967\": rpc error: code = NotFound desc = could not find container \"0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967\": container with ID starting with 0df07a13e050eaa14d2abf4d79fba90a1aefb1cf737654e55e1e21f9d8a8e967 not found: ID does not exist" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.479902 4692 scope.go:117] "RemoveContainer" containerID="19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769" Mar 09 09:23:34 crc kubenswrapper[4692]: E0309 09:23:34.480185 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\": container with ID starting with 19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769 not found: ID does not exist" containerID="19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769" Mar 09 09:23:34 crc kubenswrapper[4692]: I0309 09:23:34.480219 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769"} err="failed to get container status \"19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\": rpc error: code = NotFound desc = could not find container \"19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769\": container with ID starting with 19f37eb657bd996cb96126b2b6d9b8630f3a830e3d0c1524cad8d4137c84f769 not found: ID does not exist" Mar 09 09:23:36 crc kubenswrapper[4692]: I0309 09:23:36.083236 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 09 09:23:36 crc kubenswrapper[4692]: I0309 09:23:36.911675 4692 status_manager.go:851] "Failed to get status for pod" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-665bb7464f-hq9xx\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:36 crc kubenswrapper[4692]: I0309 09:23:36.912063 4692 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:36 crc kubenswrapper[4692]: I0309 09:23:36.912987 4692 status_manager.go:851] "Failed to get status for pod" podUID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:36 crc kubenswrapper[4692]: I0309 09:23:36.913447 4692 status_manager.go:851] "Failed to get status for pod" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-665bb7464f-hq9xx\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:36 crc kubenswrapper[4692]: I0309 09:23:36.913719 4692 status_manager.go:851] "Failed to get status for pod" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-87c746796-wbqnz\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:39 crc kubenswrapper[4692]: E0309 09:23:39.630597 4692 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:39 crc kubenswrapper[4692]: E0309 09:23:39.631360 4692 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:39 crc kubenswrapper[4692]: E0309 09:23:39.631808 4692 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:39 crc kubenswrapper[4692]: E0309 09:23:39.632283 4692 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:39 crc kubenswrapper[4692]: E0309 09:23:39.632594 4692 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:39 crc kubenswrapper[4692]: I0309 09:23:39.632634 4692 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 09 09:23:39 crc kubenswrapper[4692]: E0309 09:23:39.632961 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="200ms" Mar 09 09:23:39 crc kubenswrapper[4692]: E0309 09:23:39.834150 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="400ms" Mar 09 09:23:40 crc kubenswrapper[4692]: I0309 09:23:40.076475 4692 status_manager.go:851] "Failed to get status for pod" podUID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:40 crc kubenswrapper[4692]: I0309 09:23:40.076887 4692 status_manager.go:851] "Failed to get status for pod" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-665bb7464f-hq9xx\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:40 crc kubenswrapper[4692]: I0309 09:23:40.077310 4692 status_manager.go:851] "Failed to get status for pod" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-87c746796-wbqnz\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:40 crc kubenswrapper[4692]: E0309 09:23:40.234778 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="800ms" Mar 09 09:23:41 crc kubenswrapper[4692]: E0309 09:23:41.036321 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="1.6s" Mar 09 09:23:41 crc kubenswrapper[4692]: E0309 09:23:41.322499 4692 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.181:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189b21f47ef51a9f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-09 09:23:32.286151327 +0000 UTC m=+213.110886908,LastTimestamp:2026-03-09 09:23:32.286151327 +0000 UTC m=+213.110886908,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 09 09:23:42 crc kubenswrapper[4692]: E0309 09:23:42.637406 4692 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.181:6443: connect: connection refused" interval="3.2s" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.070762 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.071626 4692 status_manager.go:851] "Failed to get status for pod" podUID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.071902 4692 status_manager.go:851] "Failed to get status for pod" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-665bb7464f-hq9xx\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.072341 4692 status_manager.go:851] "Failed to get status for pod" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-87c746796-wbqnz\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.091429 4692 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.091648 4692 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:43 crc kubenswrapper[4692]: E0309 09:23:43.092102 4692 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.092852 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.455545 4692 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="52808002d9fdbe43b9db157b2d6266d8715d082b51b8e1a777106060a4488ab0" exitCode=0 Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.455586 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"52808002d9fdbe43b9db157b2d6266d8715d082b51b8e1a777106060a4488ab0"} Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.455616 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f6c2b259b4d0ca9883364c5abe4d171b8983cf1edef2fb48d7276fb3c8585890"} Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.455998 4692 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.456030 4692 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.456353 4692 status_manager.go:851] "Failed to get status for pod" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" pod="openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-665bb7464f-hq9xx\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:43 crc kubenswrapper[4692]: E0309 09:23:43.456421 4692 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.456711 4692 status_manager.go:851] "Failed to get status for pod" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" pod="openshift-controller-manager/controller-manager-87c746796-wbqnz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-87c746796-wbqnz\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:43 crc kubenswrapper[4692]: I0309 09:23:43.457037 4692 status_manager.go:851] "Failed to get status for pod" podUID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.181:6443: connect: connection refused" Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.461512 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2c08213dadaa7cc8ab0376ee03d125bcbcd5afe8d563a5b683ddb29c26d8b21c"} Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.461883 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"21cf2cf9fba8aa9ec68c803b354f8435e5e875a5b47e632cbdfa4a542f167969"} Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.461899 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"552aeb42dff12aab15b4189af25e4b7acf27822655be73860b58ea115b932ae4"} Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.461911 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"125208d1a82005631a03a94ddae9be1764327bcfbe1154925da8e04d7c4d1852"} Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.461921 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6afb33464489b94131383483ef75e4903693ae9462eda54d6d304bfa0c3ebb31"} Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.462243 4692 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.462261 4692 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:44 crc kubenswrapper[4692]: I0309 09:23:44.462564 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:45 crc kubenswrapper[4692]: I0309 09:23:45.799895 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" podUID="5537a456-f771-4a55-b811-f20b50cd6446" containerName="oauth-openshift" containerID="cri-o://6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff" gracePeriod=15 Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.214919 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343487 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-login\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343541 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-ocp-branding-template\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343573 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-trusted-ca-bundle\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343588 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-error\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343633 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-session\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343652 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5537a456-f771-4a55-b811-f20b50cd6446-audit-dir\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343686 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-provider-selection\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343713 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-service-ca\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343735 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-audit-policies\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343753 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-idp-0-file-data\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343788 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwlq\" (UniqueName: \"kubernetes.io/projected/5537a456-f771-4a55-b811-f20b50cd6446-kube-api-access-kfwlq\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343807 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-router-certs\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343823 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-cliconfig\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.343842 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-serving-cert\") pod \"5537a456-f771-4a55-b811-f20b50cd6446\" (UID: \"5537a456-f771-4a55-b811-f20b50cd6446\") " Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.345033 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.345403 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.345517 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.345679 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.346068 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5537a456-f771-4a55-b811-f20b50cd6446-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.351345 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.353043 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.353847 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.354253 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.354439 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.355285 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.356100 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.356248 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5537a456-f771-4a55-b811-f20b50cd6446-kube-api-access-kfwlq" (OuterVolumeSpecName: "kube-api-access-kfwlq") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "kube-api-access-kfwlq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.358608 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "5537a456-f771-4a55-b811-f20b50cd6446" (UID: "5537a456-f771-4a55-b811-f20b50cd6446"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445431 4692 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445481 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445499 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwlq\" (UniqueName: \"kubernetes.io/projected/5537a456-f771-4a55-b811-f20b50cd6446-kube-api-access-kfwlq\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445514 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445529 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445542 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445555 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445570 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445583 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445597 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445615 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445629 4692 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5537a456-f771-4a55-b811-f20b50cd6446-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445643 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.445657 4692 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5537a456-f771-4a55-b811-f20b50cd6446-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.472014 4692 generic.go:334] "Generic (PLEG): container finished" podID="5537a456-f771-4a55-b811-f20b50cd6446" containerID="6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff" exitCode=0 Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.472055 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" event={"ID":"5537a456-f771-4a55-b811-f20b50cd6446","Type":"ContainerDied","Data":"6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff"} Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.472079 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" event={"ID":"5537a456-f771-4a55-b811-f20b50cd6446","Type":"ContainerDied","Data":"0f688c61781d157cd9f4d3521383f882b2391404f59f65310fab48206e98737b"} Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.472093 4692 scope.go:117] "RemoveContainer" containerID="6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.472201 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-pbl6c" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.493046 4692 scope.go:117] "RemoveContainer" containerID="6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff" Mar 09 09:23:46 crc kubenswrapper[4692]: E0309 09:23:46.493543 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff\": container with ID starting with 6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff not found: ID does not exist" containerID="6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff" Mar 09 09:23:46 crc kubenswrapper[4692]: I0309 09:23:46.493581 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff"} err="failed to get container status \"6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff\": rpc error: code = NotFound desc = could not find container \"6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff\": container with ID starting with 6276446ee8729dd627e11c5a35bd52ec9de1d67b6f00b6415204039a9bcf42ff not found: ID does not exist" Mar 09 09:23:47 crc kubenswrapper[4692]: I0309 09:23:47.478946 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 09:23:47 crc kubenswrapper[4692]: I0309 09:23:47.480573 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 09 09:23:47 crc kubenswrapper[4692]: I0309 09:23:47.480671 4692 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0c021cb088029d8da4b3a5fee28a92cd55b25ceb658789d94bf42b674993f985" exitCode=1 Mar 09 09:23:47 crc kubenswrapper[4692]: I0309 09:23:47.480746 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0c021cb088029d8da4b3a5fee28a92cd55b25ceb658789d94bf42b674993f985"} Mar 09 09:23:47 crc kubenswrapper[4692]: I0309 09:23:47.481819 4692 scope.go:117] "RemoveContainer" containerID="0c021cb088029d8da4b3a5fee28a92cd55b25ceb658789d94bf42b674993f985" Mar 09 09:23:47 crc kubenswrapper[4692]: I0309 09:23:47.623955 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:23:47 crc kubenswrapper[4692]: I0309 09:23:47.624023 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:23:48 crc kubenswrapper[4692]: I0309 09:23:48.093897 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:48 crc kubenswrapper[4692]: I0309 09:23:48.093936 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:48 crc kubenswrapper[4692]: I0309 09:23:48.098389 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:48 crc kubenswrapper[4692]: I0309 09:23:48.488474 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 09:23:48 crc kubenswrapper[4692]: I0309 09:23:48.489822 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 09 09:23:48 crc kubenswrapper[4692]: I0309 09:23:48.489879 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"83c1b86c1ca17a9fbf1702ef3fe00c474860d354e2120c2e99d409e161ef6d5e"} Mar 09 09:23:50 crc kubenswrapper[4692]: I0309 09:23:50.159447 4692 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:50 crc kubenswrapper[4692]: I0309 09:23:50.335087 4692 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="790c7403-e2b0-44b0-87e7-9eac8c4413c9" Mar 09 09:23:50 crc kubenswrapper[4692]: I0309 09:23:50.501446 4692 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:50 crc kubenswrapper[4692]: I0309 09:23:50.501482 4692 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:50 crc kubenswrapper[4692]: I0309 09:23:50.504860 4692 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="790c7403-e2b0-44b0-87e7-9eac8c4413c9" Mar 09 09:23:50 crc kubenswrapper[4692]: I0309 09:23:50.506310 4692 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://6afb33464489b94131383483ef75e4903693ae9462eda54d6d304bfa0c3ebb31" Mar 09 09:23:50 crc kubenswrapper[4692]: I0309 09:23:50.506367 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:23:51 crc kubenswrapper[4692]: I0309 09:23:51.506649 4692 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:51 crc kubenswrapper[4692]: I0309 09:23:51.506675 4692 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ebe078e2-7773-4e2e-b1d8-37078e375a7d" Mar 09 09:23:51 crc kubenswrapper[4692]: I0309 09:23:51.510526 4692 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="790c7403-e2b0-44b0-87e7-9eac8c4413c9" Mar 09 09:23:53 crc kubenswrapper[4692]: I0309 09:23:53.794493 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:23:56 crc kubenswrapper[4692]: I0309 09:23:56.151137 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:23:56 crc kubenswrapper[4692]: I0309 09:23:56.151471 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 09 09:23:56 crc kubenswrapper[4692]: I0309 09:23:56.151512 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 09 09:23:56 crc kubenswrapper[4692]: I0309 09:23:56.748364 4692 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 09 09:23:56 crc kubenswrapper[4692]: I0309 09:23:56.782562 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 09 09:23:57 crc kubenswrapper[4692]: I0309 09:23:57.083769 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 09 09:23:57 crc kubenswrapper[4692]: I0309 09:23:57.397492 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 09 09:23:58 crc kubenswrapper[4692]: I0309 09:23:58.504973 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 09 09:23:58 crc kubenswrapper[4692]: I0309 09:23:58.936077 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 09 09:24:00 crc kubenswrapper[4692]: I0309 09:24:00.146395 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 09 09:24:00 crc kubenswrapper[4692]: I0309 09:24:00.664151 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 09 09:24:00 crc kubenswrapper[4692]: I0309 09:24:00.685833 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 09 09:24:01 crc kubenswrapper[4692]: I0309 09:24:01.066506 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 09 09:24:01 crc kubenswrapper[4692]: I0309 09:24:01.140611 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 09 09:24:01 crc kubenswrapper[4692]: I0309 09:24:01.266007 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 09 09:24:01 crc kubenswrapper[4692]: I0309 09:24:01.456235 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 09 09:24:01 crc kubenswrapper[4692]: I0309 09:24:01.828820 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 09 09:24:02 crc kubenswrapper[4692]: I0309 09:24:02.033698 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 09 09:24:02 crc kubenswrapper[4692]: I0309 09:24:02.155703 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 09 09:24:02 crc kubenswrapper[4692]: I0309 09:24:02.742051 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 09 09:24:02 crc kubenswrapper[4692]: I0309 09:24:02.811084 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 09 09:24:02 crc kubenswrapper[4692]: I0309 09:24:02.876884 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 09 09:24:03 crc kubenswrapper[4692]: I0309 09:24:03.515824 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 09 09:24:03 crc kubenswrapper[4692]: I0309 09:24:03.600372 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 09 09:24:03 crc kubenswrapper[4692]: I0309 09:24:03.718122 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.141052 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.264552 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.459995 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.482588 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.494370 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.533358 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.550551 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.558016 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.614990 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.646145 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.713659 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.905715 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 09 09:24:04 crc kubenswrapper[4692]: I0309 09:24:04.914377 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.032248 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.079912 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.261302 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.268663 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.280137 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.285149 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.454944 4692 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.458913 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-87c746796-wbqnz","openshift-route-controller-manager/route-controller-manager-665bb7464f-hq9xx","openshift-authentication/oauth-openshift-558db77b4-pbl6c","openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.458973 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.463726 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.478691 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.478666744 podStartE2EDuration="15.478666744s" podCreationTimestamp="2026-03-09 09:23:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:24:05.474805575 +0000 UTC m=+246.299541176" watchObservedRunningTime="2026-03-09 09:24:05.478666744 +0000 UTC m=+246.303402355" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.526966 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.564602 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.567636 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.586138 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.629200 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.769943 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.771392 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.834149 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.863204 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.870923 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.960668 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 09 09:24:05 crc kubenswrapper[4692]: I0309 09:24:05.971314 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.017962 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.019157 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.077751 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" path="/var/lib/kubelet/pods/2d0ec70a-c384-4fe0-ae86-3e0da3f3637e/volumes" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.078406 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" path="/var/lib/kubelet/pods/4ad5398b-031a-4fe6-a264-9e0df09ade1a/volumes" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.078914 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5537a456-f771-4a55-b811-f20b50cd6446" path="/var/lib/kubelet/pods/5537a456-f771-4a55-b811-f20b50cd6446/volumes" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.120772 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.151248 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.151300 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.584717 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.641053 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.780972 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.785751 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 09 09:24:06 crc kubenswrapper[4692]: I0309 09:24:06.919083 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.016837 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.083071 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.102410 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.148247 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.234905 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.298001 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.303665 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.387643 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.435292 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.447719 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.475397 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.525189 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.609067 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.687742 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.786061 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.791584 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.810611 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.825502 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 09 09:24:07 crc kubenswrapper[4692]: I0309 09:24:07.957601 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.039277 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.076083 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.148670 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.199947 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.310374 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.384713 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.391456 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.541630 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.615617 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.622162 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.645191 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.733281 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.793604 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 09 09:24:08 crc kubenswrapper[4692]: I0309 09:24:08.994043 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.018960 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.130800 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.234250 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.241446 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.305023 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.360998 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.372120 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.416823 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.472791 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.555962 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.567674 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.637999 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.788555 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.870398 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.905908 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.949540 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 09 09:24:09 crc kubenswrapper[4692]: I0309 09:24:09.949617 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.194670 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.216865 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.276715 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.296152 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.349633 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.443575 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.457637 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.630289 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.666330 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.692322 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.863208 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.863217 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.902897 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.908396 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.931191 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.935626 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.958756 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 09 09:24:10 crc kubenswrapper[4692]: I0309 09:24:10.979320 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.011423 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.109087 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.200254 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.290231 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.300253 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.316280 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.399299 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.447023 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.555876 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.573842 4692 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.584927 4692 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.585212 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://9f2e4096577478b98ac14cd87c69a2d5b8e851ac6682e739bae9f5c7446275a2" gracePeriod=5 Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.611975 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.807539 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 09 09:24:11 crc kubenswrapper[4692]: I0309 09:24:11.946944 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.010530 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.048933 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.052296 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.066429 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.108439 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.140405 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.311506 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.333465 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.416189 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.484102 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.484181 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.517416 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.542801 4692 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.632377 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.652887 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.685422 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 09 09:24:12 crc kubenswrapper[4692]: I0309 09:24:12.821511 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.008518 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.087825 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.119971 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.172870 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.315889 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.401880 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.454104 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.508143 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.629401 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.661931 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.710851 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.749638 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 09 09:24:13 crc kubenswrapper[4692]: I0309 09:24:13.776753 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.026324 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.047710 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.172689 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.223632 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.254782 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.349983 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.352804 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.446810 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.546634 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.555779 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.619057 4692 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.844491 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.878406 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.887246 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 09 09:24:14 crc kubenswrapper[4692]: I0309 09:24:14.933771 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.047852 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.101648 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.257305 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.394674 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.555381 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.658519 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.714737 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.788762 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.847202 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.861720 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.894769 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.913863 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 09 09:24:15 crc kubenswrapper[4692]: I0309 09:24:15.994251 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.005812 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.027912 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.151404 4692 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.151466 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.151518 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.152181 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"83c1b86c1ca17a9fbf1702ef3fe00c474860d354e2120c2e99d409e161ef6d5e"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.152311 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://83c1b86c1ca17a9fbf1702ef3fe00c474860d354e2120c2e99d409e161ef6d5e" gracePeriod=30 Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.208375 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.305939 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.462416 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.648263 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 09 09:24:16 crc kubenswrapper[4692]: I0309 09:24:16.648325 4692 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="9f2e4096577478b98ac14cd87c69a2d5b8e851ac6682e739bae9f5c7446275a2" exitCode=137 Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.054331 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.094511 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.185650 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.185766 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.199312 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265182 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75f74655f4-kdsps"] Mar 09 09:24:17 crc kubenswrapper[4692]: E0309 09:24:17.265607 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" containerName="controller-manager" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265634 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" containerName="controller-manager" Mar 09 09:24:17 crc kubenswrapper[4692]: E0309 09:24:17.265656 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5537a456-f771-4a55-b811-f20b50cd6446" containerName="oauth-openshift" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265669 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5537a456-f771-4a55-b811-f20b50cd6446" containerName="oauth-openshift" Mar 09 09:24:17 crc kubenswrapper[4692]: E0309 09:24:17.265694 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" containerName="installer" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265705 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" containerName="installer" Mar 09 09:24:17 crc kubenswrapper[4692]: E0309 09:24:17.265720 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" containerName="route-controller-manager" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265730 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" containerName="route-controller-manager" Mar 09 09:24:17 crc kubenswrapper[4692]: E0309 09:24:17.265751 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265761 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265928 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265951 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad5398b-031a-4fe6-a264-9e0df09ade1a" containerName="route-controller-manager" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265967 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61e177f-a1d1-4a8b-a7d7-24035a88f54e" containerName="installer" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.265984 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5537a456-f771-4a55-b811-f20b50cd6446" containerName="oauth-openshift" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.266000 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d0ec70a-c384-4fe0-ae86-3e0da3f3637e" containerName="controller-manager" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.266603 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.266690 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.271018 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.271158 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.271444 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.271504 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.271608 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.272553 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.275484 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh"] Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.276595 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.279352 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.279803 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.279844 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.280150 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.280362 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-66d4b9c857-7szfz"] Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.280579 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.280716 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.281019 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.281854 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.284109 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75f74655f4-kdsps"] Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.286031 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh"] Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.286763 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.287422 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.287912 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.288017 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.288071 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.288086 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.289532 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.290784 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.290931 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.291551 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.291674 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.291860 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.296283 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.307316 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.324146 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.324505 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.324600 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.324860 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.325210 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.325421 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.325831 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.325858 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.326053 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.331710 4692 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.331734 4692 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.331811 4692 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.331829 4692 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.341066 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.342991 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.343026 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66d4b9c857-7szfz"] Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.344278 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.383346 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432809 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-login\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432870 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432898 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb67e7b-248e-43ba-8eef-02e696de148a-serving-cert\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432916 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-error\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432934 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-session\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432957 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twxcf\" (UniqueName: \"kubernetes.io/projected/4873ebc1-4086-48a8-82f5-331a1a5f47b3-kube-api-access-twxcf\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432979 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-config\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.432998 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433017 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433041 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433057 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gttxr\" (UniqueName: \"kubernetes.io/projected/4200c76f-d715-479c-a4fe-113b6024c625-kube-api-access-gttxr\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433074 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4873ebc1-4086-48a8-82f5-331a1a5f47b3-config\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433094 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-audit-policies\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433110 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4200c76f-d715-479c-a4fe-113b6024c625-audit-dir\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433127 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-service-ca\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433145 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-proxy-ca-bundles\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433182 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4873ebc1-4086-48a8-82f5-331a1a5f47b3-client-ca\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433206 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433228 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433244 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6d2d\" (UniqueName: \"kubernetes.io/projected/5bb67e7b-248e-43ba-8eef-02e696de148a-kube-api-access-b6d2d\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433260 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-router-certs\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433277 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-client-ca\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433295 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4873ebc1-4086-48a8-82f5-331a1a5f47b3-serving-cert\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.433325 4692 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.461850 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.534780 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-audit-policies\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535469 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4200c76f-d715-479c-a4fe-113b6024c625-audit-dir\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535576 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-service-ca\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535613 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4200c76f-d715-479c-a4fe-113b6024c625-audit-dir\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535623 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-proxy-ca-bundles\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535694 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-audit-policies\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535735 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535763 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4873ebc1-4086-48a8-82f5-331a1a5f47b3-client-ca\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535789 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535812 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6d2d\" (UniqueName: \"kubernetes.io/projected/5bb67e7b-248e-43ba-8eef-02e696de148a-kube-api-access-b6d2d\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535832 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-router-certs\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535862 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-client-ca\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535902 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4873ebc1-4086-48a8-82f5-331a1a5f47b3-serving-cert\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535940 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-login\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535957 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.535987 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb67e7b-248e-43ba-8eef-02e696de148a-serving-cert\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536008 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-error\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536031 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-session\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536055 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twxcf\" (UniqueName: \"kubernetes.io/projected/4873ebc1-4086-48a8-82f5-331a1a5f47b3-kube-api-access-twxcf\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536081 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-config\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536104 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536124 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536186 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536207 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gttxr\" (UniqueName: \"kubernetes.io/projected/4200c76f-d715-479c-a4fe-113b6024c625-kube-api-access-gttxr\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536231 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4873ebc1-4086-48a8-82f5-331a1a5f47b3-config\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536387 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-service-ca\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536496 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.536746 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-proxy-ca-bundles\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.537360 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4873ebc1-4086-48a8-82f5-331a1a5f47b3-client-ca\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.537516 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-client-ca\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.537548 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4873ebc1-4086-48a8-82f5-331a1a5f47b3-config\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.538729 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb67e7b-248e-43ba-8eef-02e696de148a-config\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.539849 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-session\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.539925 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.539933 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb67e7b-248e-43ba-8eef-02e696de148a-serving-cert\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.540520 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.540592 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-login\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.541508 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.542636 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-template-error\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.543036 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.544719 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.545031 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4873ebc1-4086-48a8-82f5-331a1a5f47b3-serving-cert\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.555736 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/4200c76f-d715-479c-a4fe-113b6024c625-v4-0-config-system-router-certs\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.560707 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6d2d\" (UniqueName: \"kubernetes.io/projected/5bb67e7b-248e-43ba-8eef-02e696de148a-kube-api-access-b6d2d\") pod \"controller-manager-75f74655f4-kdsps\" (UID: \"5bb67e7b-248e-43ba-8eef-02e696de148a\") " pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.561654 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twxcf\" (UniqueName: \"kubernetes.io/projected/4873ebc1-4086-48a8-82f5-331a1a5f47b3-kube-api-access-twxcf\") pod \"route-controller-manager-7ffff98dff-gbtgh\" (UID: \"4873ebc1-4086-48a8-82f5-331a1a5f47b3\") " pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.562280 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gttxr\" (UniqueName: \"kubernetes.io/projected/4200c76f-d715-479c-a4fe-113b6024c625-kube-api-access-gttxr\") pod \"oauth-openshift-66d4b9c857-7szfz\" (UID: \"4200c76f-d715-479c-a4fe-113b6024c625\") " pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.601178 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.601382 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.602766 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.613566 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.624527 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.624679 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.625798 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.665456 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.665597 4692 scope.go:117] "RemoveContainer" containerID="9f2e4096577478b98ac14cd87c69a2d5b8e851ac6682e739bae9f5c7446275a2" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.665820 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 09 09:24:17 crc kubenswrapper[4692]: I0309 09:24:17.922546 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.045182 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh"] Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.084711 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.113426 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75f74655f4-kdsps"] Mar 09 09:24:18 crc kubenswrapper[4692]: W0309 09:24:18.126748 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bb67e7b_248e_43ba_8eef_02e696de148a.slice/crio-aaed5867b475727c81460ec6478110afdff059406a25cd32784dca2086b893b8 WatchSource:0}: Error finding container aaed5867b475727c81460ec6478110afdff059406a25cd32784dca2086b893b8: Status 404 returned error can't find the container with id aaed5867b475727c81460ec6478110afdff059406a25cd32784dca2086b893b8 Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.138994 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66d4b9c857-7szfz"] Mar 09 09:24:18 crc kubenswrapper[4692]: W0309 09:24:18.151922 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4200c76f_d715_479c_a4fe_113b6024c625.slice/crio-fb8bf116a2853ca741a6e58244b88519fd3e67890621f05dba8be147f0ee0754 WatchSource:0}: Error finding container fb8bf116a2853ca741a6e58244b88519fd3e67890621f05dba8be147f0ee0754: Status 404 returned error can't find the container with id fb8bf116a2853ca741a6e58244b88519fd3e67890621f05dba8be147f0ee0754 Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.208580 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.631753 4692 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.657405 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.672885 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" event={"ID":"4873ebc1-4086-48a8-82f5-331a1a5f47b3","Type":"ContainerStarted","Data":"6c86f9cda9aecd43f7f1a9d8ac280aab5e0f26a436eb3f7022a3dcf9a5040e89"} Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.672951 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" event={"ID":"4873ebc1-4086-48a8-82f5-331a1a5f47b3","Type":"ContainerStarted","Data":"cedd833b3a2d07e35e00044b3ac4ccf18393a64e9aada3c4e22977ab10a9e8c7"} Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.673093 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.676878 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" event={"ID":"5bb67e7b-248e-43ba-8eef-02e696de148a","Type":"ContainerStarted","Data":"78eb842c4bc58e32561bab4b58ab5b578e172fd33070422897e1c89314581f2f"} Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.676918 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" event={"ID":"5bb67e7b-248e-43ba-8eef-02e696de148a","Type":"ContainerStarted","Data":"aaed5867b475727c81460ec6478110afdff059406a25cd32784dca2086b893b8"} Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.678307 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.679750 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" event={"ID":"4200c76f-d715-479c-a4fe-113b6024c625","Type":"ContainerStarted","Data":"2f2f9a55359b75f3cf7b7680b783d595f6cb5d5183e78afc05d68969c7b12a26"} Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.679776 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" event={"ID":"4200c76f-d715-479c-a4fe-113b6024c625","Type":"ContainerStarted","Data":"fb8bf116a2853ca741a6e58244b88519fd3e67890621f05dba8be147f0ee0754"} Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.679967 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.684059 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.702920 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" podStartSLOduration=47.702900661 podStartE2EDuration="47.702900661s" podCreationTimestamp="2026-03-09 09:23:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:24:18.70219741 +0000 UTC m=+259.526932981" watchObservedRunningTime="2026-03-09 09:24:18.702900661 +0000 UTC m=+259.527636242" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.708906 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7ffff98dff-gbtgh" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.747421 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75f74655f4-kdsps" podStartSLOduration=47.747399174 podStartE2EDuration="47.747399174s" podCreationTimestamp="2026-03-09 09:23:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:24:18.72703528 +0000 UTC m=+259.551770871" watchObservedRunningTime="2026-03-09 09:24:18.747399174 +0000 UTC m=+259.572134755" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.784319 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" podStartSLOduration=58.784296334 podStartE2EDuration="58.784296334s" podCreationTimestamp="2026-03-09 09:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:24:18.74924835 +0000 UTC m=+259.573983931" watchObservedRunningTime="2026-03-09 09:24:18.784296334 +0000 UTC m=+259.609031915" Mar 09 09:24:18 crc kubenswrapper[4692]: I0309 09:24:18.830886 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 09 09:24:19 crc kubenswrapper[4692]: I0309 09:24:19.026681 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-66d4b9c857-7szfz" Mar 09 09:24:19 crc kubenswrapper[4692]: I0309 09:24:19.300526 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 09 09:24:20 crc kubenswrapper[4692]: I0309 09:24:20.798931 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 09 09:24:20 crc kubenswrapper[4692]: I0309 09:24:20.915942 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 09 09:24:21 crc kubenswrapper[4692]: I0309 09:24:21.047079 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 09 09:24:40 crc kubenswrapper[4692]: I0309 09:24:40.810745 4692 generic.go:334] "Generic (PLEG): container finished" podID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerID="d7247aed919cff304a8b476a12e389279bfab9c4696fa35749265c5d9f1f327d" exitCode=0 Mar 09 09:24:40 crc kubenswrapper[4692]: I0309 09:24:40.810845 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" event={"ID":"96013f1f-d2b0-42ec-a4f3-3626f48be8e1","Type":"ContainerDied","Data":"d7247aed919cff304a8b476a12e389279bfab9c4696fa35749265c5d9f1f327d"} Mar 09 09:24:40 crc kubenswrapper[4692]: I0309 09:24:40.811645 4692 scope.go:117] "RemoveContainer" containerID="d7247aed919cff304a8b476a12e389279bfab9c4696fa35749265c5d9f1f327d" Mar 09 09:24:41 crc kubenswrapper[4692]: I0309 09:24:41.817642 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" event={"ID":"96013f1f-d2b0-42ec-a4f3-3626f48be8e1","Type":"ContainerStarted","Data":"4320322bc11ceae70df4eb69bbb20e01c609f6a78877297e606e084168324a84"} Mar 09 09:24:41 crc kubenswrapper[4692]: I0309 09:24:41.818482 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:24:41 crc kubenswrapper[4692]: I0309 09:24:41.820461 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:24:46 crc kubenswrapper[4692]: I0309 09:24:46.852152 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 09 09:24:46 crc kubenswrapper[4692]: I0309 09:24:46.853458 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 09:24:46 crc kubenswrapper[4692]: I0309 09:24:46.854924 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 09 09:24:46 crc kubenswrapper[4692]: I0309 09:24:46.854981 4692 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="83c1b86c1ca17a9fbf1702ef3fe00c474860d354e2120c2e99d409e161ef6d5e" exitCode=137 Mar 09 09:24:46 crc kubenswrapper[4692]: I0309 09:24:46.855029 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"83c1b86c1ca17a9fbf1702ef3fe00c474860d354e2120c2e99d409e161ef6d5e"} Mar 09 09:24:46 crc kubenswrapper[4692]: I0309 09:24:46.855081 4692 scope.go:117] "RemoveContainer" containerID="0c021cb088029d8da4b3a5fee28a92cd55b25ceb658789d94bf42b674993f985" Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.624060 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.624387 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.624473 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.624992 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b30ec1f14935ee72de8d696ab21516444d1dbed8e188660550c2b615537f1de4"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.625953 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://b30ec1f14935ee72de8d696ab21516444d1dbed8e188660550c2b615537f1de4" gracePeriod=600 Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.862583 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.863447 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/1.log" Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.865282 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e275fc627b43ff52dbb15a7af40565fbdfec8e44752dc0e88844ffa466e57346"} Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.867828 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="b30ec1f14935ee72de8d696ab21516444d1dbed8e188660550c2b615537f1de4" exitCode=0 Mar 09 09:24:47 crc kubenswrapper[4692]: I0309 09:24:47.867890 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"b30ec1f14935ee72de8d696ab21516444d1dbed8e188660550c2b615537f1de4"} Mar 09 09:24:48 crc kubenswrapper[4692]: I0309 09:24:48.876704 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"61fae0ebf34e8ab848fb0659a787a8b14e18310fadde50c12777c1d7eafc19fe"} Mar 09 09:24:53 crc kubenswrapper[4692]: I0309 09:24:53.794869 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:24:56 crc kubenswrapper[4692]: I0309 09:24:56.150871 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:24:56 crc kubenswrapper[4692]: I0309 09:24:56.154995 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:24:56 crc kubenswrapper[4692]: I0309 09:24:56.926110 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 09 09:25:03 crc kubenswrapper[4692]: I0309 09:25:03.791452 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550804-bcnrd"] Mar 09 09:25:03 crc kubenswrapper[4692]: I0309 09:25:03.792565 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550804-bcnrd" Mar 09 09:25:03 crc kubenswrapper[4692]: I0309 09:25:03.794705 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:25:03 crc kubenswrapper[4692]: I0309 09:25:03.794923 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:25:03 crc kubenswrapper[4692]: I0309 09:25:03.805494 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:25:03 crc kubenswrapper[4692]: I0309 09:25:03.812291 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550804-bcnrd"] Mar 09 09:25:03 crc kubenswrapper[4692]: I0309 09:25:03.966859 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f68pn\" (UniqueName: \"kubernetes.io/projected/c1e0e894-d531-4ca0-a408-1118c2fa993b-kube-api-access-f68pn\") pod \"auto-csr-approver-29550804-bcnrd\" (UID: \"c1e0e894-d531-4ca0-a408-1118c2fa993b\") " pod="openshift-infra/auto-csr-approver-29550804-bcnrd" Mar 09 09:25:04 crc kubenswrapper[4692]: I0309 09:25:04.068245 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f68pn\" (UniqueName: \"kubernetes.io/projected/c1e0e894-d531-4ca0-a408-1118c2fa993b-kube-api-access-f68pn\") pod \"auto-csr-approver-29550804-bcnrd\" (UID: \"c1e0e894-d531-4ca0-a408-1118c2fa993b\") " pod="openshift-infra/auto-csr-approver-29550804-bcnrd" Mar 09 09:25:04 crc kubenswrapper[4692]: I0309 09:25:04.088083 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f68pn\" (UniqueName: \"kubernetes.io/projected/c1e0e894-d531-4ca0-a408-1118c2fa993b-kube-api-access-f68pn\") pod \"auto-csr-approver-29550804-bcnrd\" (UID: \"c1e0e894-d531-4ca0-a408-1118c2fa993b\") " pod="openshift-infra/auto-csr-approver-29550804-bcnrd" Mar 09 09:25:04 crc kubenswrapper[4692]: I0309 09:25:04.109035 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550804-bcnrd" Mar 09 09:25:04 crc kubenswrapper[4692]: I0309 09:25:04.498438 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550804-bcnrd"] Mar 09 09:25:04 crc kubenswrapper[4692]: I0309 09:25:04.961255 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550804-bcnrd" event={"ID":"c1e0e894-d531-4ca0-a408-1118c2fa993b","Type":"ContainerStarted","Data":"5484130c3e967e4d2b0256fb66f2ea775e1f3a4550c7554f0cb58f62b005e641"} Mar 09 09:25:05 crc kubenswrapper[4692]: I0309 09:25:05.967413 4692 generic.go:334] "Generic (PLEG): container finished" podID="c1e0e894-d531-4ca0-a408-1118c2fa993b" containerID="714119e2a48d8ee67c9b24a147de76be865fafb71e6d7e25d7d174f99ab0974e" exitCode=0 Mar 09 09:25:05 crc kubenswrapper[4692]: I0309 09:25:05.967471 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550804-bcnrd" event={"ID":"c1e0e894-d531-4ca0-a408-1118c2fa993b","Type":"ContainerDied","Data":"714119e2a48d8ee67c9b24a147de76be865fafb71e6d7e25d7d174f99ab0974e"} Mar 09 09:25:07 crc kubenswrapper[4692]: I0309 09:25:07.227799 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550804-bcnrd" Mar 09 09:25:07 crc kubenswrapper[4692]: I0309 09:25:07.410948 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f68pn\" (UniqueName: \"kubernetes.io/projected/c1e0e894-d531-4ca0-a408-1118c2fa993b-kube-api-access-f68pn\") pod \"c1e0e894-d531-4ca0-a408-1118c2fa993b\" (UID: \"c1e0e894-d531-4ca0-a408-1118c2fa993b\") " Mar 09 09:25:07 crc kubenswrapper[4692]: I0309 09:25:07.416563 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1e0e894-d531-4ca0-a408-1118c2fa993b-kube-api-access-f68pn" (OuterVolumeSpecName: "kube-api-access-f68pn") pod "c1e0e894-d531-4ca0-a408-1118c2fa993b" (UID: "c1e0e894-d531-4ca0-a408-1118c2fa993b"). InnerVolumeSpecName "kube-api-access-f68pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:25:07 crc kubenswrapper[4692]: I0309 09:25:07.512708 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f68pn\" (UniqueName: \"kubernetes.io/projected/c1e0e894-d531-4ca0-a408-1118c2fa993b-kube-api-access-f68pn\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:07 crc kubenswrapper[4692]: I0309 09:25:07.978789 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550804-bcnrd" event={"ID":"c1e0e894-d531-4ca0-a408-1118c2fa993b","Type":"ContainerDied","Data":"5484130c3e967e4d2b0256fb66f2ea775e1f3a4550c7554f0cb58f62b005e641"} Mar 09 09:25:07 crc kubenswrapper[4692]: I0309 09:25:07.979060 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5484130c3e967e4d2b0256fb66f2ea775e1f3a4550c7554f0cb58f62b005e641" Mar 09 09:25:07 crc kubenswrapper[4692]: I0309 09:25:07.978867 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550804-bcnrd" Mar 09 09:25:32 crc kubenswrapper[4692]: I0309 09:25:32.916931 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xkh5l"] Mar 09 09:25:32 crc kubenswrapper[4692]: E0309 09:25:32.917631 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e0e894-d531-4ca0-a408-1118c2fa993b" containerName="oc" Mar 09 09:25:32 crc kubenswrapper[4692]: I0309 09:25:32.917643 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e0e894-d531-4ca0-a408-1118c2fa993b" containerName="oc" Mar 09 09:25:32 crc kubenswrapper[4692]: I0309 09:25:32.917732 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e0e894-d531-4ca0-a408-1118c2fa993b" containerName="oc" Mar 09 09:25:32 crc kubenswrapper[4692]: I0309 09:25:32.918072 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:32 crc kubenswrapper[4692]: I0309 09:25:32.930115 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xkh5l"] Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033613 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a95f4eaa-e097-478e-87c3-897e3e8bd643-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033661 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-registry-tls\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033684 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-bound-sa-token\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033705 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a95f4eaa-e097-478e-87c3-897e3e8bd643-trusted-ca\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033732 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twblr\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-kube-api-access-twblr\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033769 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a95f4eaa-e097-478e-87c3-897e3e8bd643-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033816 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.033843 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a95f4eaa-e097-478e-87c3-897e3e8bd643-registry-certificates\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.090066 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.134688 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a95f4eaa-e097-478e-87c3-897e3e8bd643-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.134746 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-registry-tls\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.134769 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-bound-sa-token\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.134787 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a95f4eaa-e097-478e-87c3-897e3e8bd643-trusted-ca\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.134804 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twblr\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-kube-api-access-twblr\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.134822 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a95f4eaa-e097-478e-87c3-897e3e8bd643-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.134859 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a95f4eaa-e097-478e-87c3-897e3e8bd643-registry-certificates\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.135375 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a95f4eaa-e097-478e-87c3-897e3e8bd643-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.136257 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a95f4eaa-e097-478e-87c3-897e3e8bd643-trusted-ca\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.136351 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a95f4eaa-e097-478e-87c3-897e3e8bd643-registry-certificates\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.140667 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a95f4eaa-e097-478e-87c3-897e3e8bd643-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.146104 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-registry-tls\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.148507 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-bound-sa-token\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.152330 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twblr\" (UniqueName: \"kubernetes.io/projected/a95f4eaa-e097-478e-87c3-897e3e8bd643-kube-api-access-twblr\") pod \"image-registry-66df7c8f76-xkh5l\" (UID: \"a95f4eaa-e097-478e-87c3-897e3e8bd643\") " pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.238820 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:33 crc kubenswrapper[4692]: I0309 09:25:33.653849 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xkh5l"] Mar 09 09:25:34 crc kubenswrapper[4692]: I0309 09:25:34.096997 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" event={"ID":"a95f4eaa-e097-478e-87c3-897e3e8bd643","Type":"ContainerStarted","Data":"20e82b24cc4262f077c1d628bd1de6f29e2054cb682081c0860917f1fa1485e8"} Mar 09 09:25:34 crc kubenswrapper[4692]: I0309 09:25:34.097549 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:34 crc kubenswrapper[4692]: I0309 09:25:34.097565 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" event={"ID":"a95f4eaa-e097-478e-87c3-897e3e8bd643","Type":"ContainerStarted","Data":"07f97ab8f237e2bdf6439684104b1e7c754a549adb181ad2f3119b32a22566d5"} Mar 09 09:25:34 crc kubenswrapper[4692]: I0309 09:25:34.118756 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" podStartSLOduration=2.118721563 podStartE2EDuration="2.118721563s" podCreationTimestamp="2026-03-09 09:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:25:34.115640159 +0000 UTC m=+334.940375750" watchObservedRunningTime="2026-03-09 09:25:34.118721563 +0000 UTC m=+334.943457154" Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.245739 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xkh5l" Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.293878 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxwb2"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.827386 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4bqhb"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.827597 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4bqhb" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="registry-server" containerID="cri-o://6ec1ecbb6a65d8760942f519a7440b6a4dcfb4e6184b9b199bc82504b9635683" gracePeriod=30 Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.836983 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ktmgp"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.837716 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ktmgp" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="registry-server" containerID="cri-o://efea09f529044bf3f09eca867fd99b9e990d01df4cf75b431b94df5261d9fb7b" gracePeriod=30 Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.848258 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdvhn"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.848507 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" containerID="cri-o://4320322bc11ceae70df4eb69bbb20e01c609f6a78877297e606e084168324a84" gracePeriod=30 Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.860068 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkkhr"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.860341 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vkkhr" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="registry-server" containerID="cri-o://cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a" gracePeriod=30 Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.860786 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pnf7m"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.861007 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pnf7m" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="registry-server" containerID="cri-o://2f11fd8acf0cbe68dc0911c7764a229e4c763847bce01b88fa085f3dd4ae6726" gracePeriod=30 Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.866514 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvcsq"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.867221 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.885898 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvcsq"] Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.918749 4692 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rdvhn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" start-of-body= Mar 09 09:25:53 crc kubenswrapper[4692]: I0309 09:25:53.918814 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.26:8080/healthz\": dial tcp 10.217.0.26:8080: connect: connection refused" Mar 09 09:25:53 crc kubenswrapper[4692]: E0309 09:25:53.995894 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a is running failed: container process not found" containerID="cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 09:25:53 crc kubenswrapper[4692]: E0309 09:25:53.996270 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a is running failed: container process not found" containerID="cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 09:25:54 crc kubenswrapper[4692]: E0309 09:25:54.003126 4692 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a is running failed: container process not found" containerID="cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a" cmd=["grpc_health_probe","-addr=:50051"] Mar 09 09:25:54 crc kubenswrapper[4692]: E0309 09:25:54.003207 4692 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-vkkhr" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="registry-server" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.018935 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrdt6\" (UniqueName: \"kubernetes.io/projected/74fc14ae-20b4-429a-8f94-886a6bc85d3e-kube-api-access-hrdt6\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.018984 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74fc14ae-20b4-429a-8f94-886a6bc85d3e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.019004 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74fc14ae-20b4-429a-8f94-886a6bc85d3e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.120547 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74fc14ae-20b4-429a-8f94-886a6bc85d3e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.120602 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74fc14ae-20b4-429a-8f94-886a6bc85d3e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.120694 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrdt6\" (UniqueName: \"kubernetes.io/projected/74fc14ae-20b4-429a-8f94-886a6bc85d3e-kube-api-access-hrdt6\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.122479 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74fc14ae-20b4-429a-8f94-886a6bc85d3e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.128291 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74fc14ae-20b4-429a-8f94-886a6bc85d3e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.137025 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrdt6\" (UniqueName: \"kubernetes.io/projected/74fc14ae-20b4-429a-8f94-886a6bc85d3e-kube-api-access-hrdt6\") pod \"marketplace-operator-79b997595-wvcsq\" (UID: \"74fc14ae-20b4-429a-8f94-886a6bc85d3e\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.204805 4692 generic.go:334] "Generic (PLEG): container finished" podID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerID="4320322bc11ceae70df4eb69bbb20e01c609f6a78877297e606e084168324a84" exitCode=0 Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.204900 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" event={"ID":"96013f1f-d2b0-42ec-a4f3-3626f48be8e1","Type":"ContainerDied","Data":"4320322bc11ceae70df4eb69bbb20e01c609f6a78877297e606e084168324a84"} Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.204941 4692 scope.go:117] "RemoveContainer" containerID="d7247aed919cff304a8b476a12e389279bfab9c4696fa35749265c5d9f1f327d" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.207607 4692 generic.go:334] "Generic (PLEG): container finished" podID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerID="2f11fd8acf0cbe68dc0911c7764a229e4c763847bce01b88fa085f3dd4ae6726" exitCode=0 Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.207654 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerDied","Data":"2f11fd8acf0cbe68dc0911c7764a229e4c763847bce01b88fa085f3dd4ae6726"} Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.211010 4692 generic.go:334] "Generic (PLEG): container finished" podID="73819651-942b-4436-b0b1-7225cfb20ef1" containerID="6ec1ecbb6a65d8760942f519a7440b6a4dcfb4e6184b9b199bc82504b9635683" exitCode=0 Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.211105 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4bqhb" event={"ID":"73819651-942b-4436-b0b1-7225cfb20ef1","Type":"ContainerDied","Data":"6ec1ecbb6a65d8760942f519a7440b6a4dcfb4e6184b9b199bc82504b9635683"} Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.211132 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4bqhb" event={"ID":"73819651-942b-4436-b0b1-7225cfb20ef1","Type":"ContainerDied","Data":"5571d809348713925825502d029a7db47086da2659bfee1591e54bf18709d5c1"} Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.211142 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5571d809348713925825502d029a7db47086da2659bfee1591e54bf18709d5c1" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.212533 4692 generic.go:334] "Generic (PLEG): container finished" podID="3d4b7447-a6e6-4530-b529-2ec461452791" containerID="cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a" exitCode=0 Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.212576 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkkhr" event={"ID":"3d4b7447-a6e6-4530-b529-2ec461452791","Type":"ContainerDied","Data":"cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a"} Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.214068 4692 generic.go:334] "Generic (PLEG): container finished" podID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerID="efea09f529044bf3f09eca867fd99b9e990d01df4cf75b431b94df5261d9fb7b" exitCode=0 Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.214093 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktmgp" event={"ID":"067264d3-8fc2-46cd-8d94-e89f9e25a500","Type":"ContainerDied","Data":"efea09f529044bf3f09eca867fd99b9e990d01df4cf75b431b94df5261d9fb7b"} Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.240126 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.246685 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.359823 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.375891 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.377234 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.385216 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.428857 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-utilities\") pod \"73819651-942b-4436-b0b1-7225cfb20ef1\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.428952 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-catalog-content\") pod \"73819651-942b-4436-b0b1-7225cfb20ef1\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.428998 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6gqq\" (UniqueName: \"kubernetes.io/projected/73819651-942b-4436-b0b1-7225cfb20ef1-kube-api-access-g6gqq\") pod \"73819651-942b-4436-b0b1-7225cfb20ef1\" (UID: \"73819651-942b-4436-b0b1-7225cfb20ef1\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.429899 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-utilities" (OuterVolumeSpecName: "utilities") pod "73819651-942b-4436-b0b1-7225cfb20ef1" (UID: "73819651-942b-4436-b0b1-7225cfb20ef1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.434479 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73819651-942b-4436-b0b1-7225cfb20ef1-kube-api-access-g6gqq" (OuterVolumeSpecName: "kube-api-access-g6gqq") pod "73819651-942b-4436-b0b1-7225cfb20ef1" (UID: "73819651-942b-4436-b0b1-7225cfb20ef1"). InnerVolumeSpecName "kube-api-access-g6gqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.490088 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73819651-942b-4436-b0b1-7225cfb20ef1" (UID: "73819651-942b-4436-b0b1-7225cfb20ef1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.527567 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvcsq"] Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532372 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-catalog-content\") pod \"067264d3-8fc2-46cd-8d94-e89f9e25a500\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532419 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics\") pod \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532442 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-catalog-content\") pod \"3d4b7447-a6e6-4530-b529-2ec461452791\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532460 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-catalog-content\") pod \"15d06ca2-19f5-4c88-8c32-51e491ffa834\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532498 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbw5p\" (UniqueName: \"kubernetes.io/projected/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-kube-api-access-sbw5p\") pod \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532541 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pnq6\" (UniqueName: \"kubernetes.io/projected/15d06ca2-19f5-4c88-8c32-51e491ffa834-kube-api-access-2pnq6\") pod \"15d06ca2-19f5-4c88-8c32-51e491ffa834\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532564 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca\") pod \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\" (UID: \"96013f1f-d2b0-42ec-a4f3-3626f48be8e1\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532582 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-utilities\") pod \"3d4b7447-a6e6-4530-b529-2ec461452791\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.532607 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gkcn\" (UniqueName: \"kubernetes.io/projected/067264d3-8fc2-46cd-8d94-e89f9e25a500-kube-api-access-6gkcn\") pod \"067264d3-8fc2-46cd-8d94-e89f9e25a500\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.535206 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "96013f1f-d2b0-42ec-a4f3-3626f48be8e1" (UID: "96013f1f-d2b0-42ec-a4f3-3626f48be8e1"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.535688 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-utilities" (OuterVolumeSpecName: "utilities") pod "3d4b7447-a6e6-4530-b529-2ec461452791" (UID: "3d4b7447-a6e6-4530-b529-2ec461452791"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.536764 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "96013f1f-d2b0-42ec-a4f3-3626f48be8e1" (UID: "96013f1f-d2b0-42ec-a4f3-3626f48be8e1"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.537459 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/067264d3-8fc2-46cd-8d94-e89f9e25a500-kube-api-access-6gkcn" (OuterVolumeSpecName: "kube-api-access-6gkcn") pod "067264d3-8fc2-46cd-8d94-e89f9e25a500" (UID: "067264d3-8fc2-46cd-8d94-e89f9e25a500"). InnerVolumeSpecName "kube-api-access-6gkcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.538376 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d06ca2-19f5-4c88-8c32-51e491ffa834-kube-api-access-2pnq6" (OuterVolumeSpecName: "kube-api-access-2pnq6") pod "15d06ca2-19f5-4c88-8c32-51e491ffa834" (UID: "15d06ca2-19f5-4c88-8c32-51e491ffa834"). InnerVolumeSpecName "kube-api-access-2pnq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.541720 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-kube-api-access-sbw5p" (OuterVolumeSpecName: "kube-api-access-sbw5p") pod "96013f1f-d2b0-42ec-a4f3-3626f48be8e1" (UID: "96013f1f-d2b0-42ec-a4f3-3626f48be8e1"). InnerVolumeSpecName "kube-api-access-sbw5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547413 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzw78\" (UniqueName: \"kubernetes.io/projected/3d4b7447-a6e6-4530-b529-2ec461452791-kube-api-access-nzw78\") pod \"3d4b7447-a6e6-4530-b529-2ec461452791\" (UID: \"3d4b7447-a6e6-4530-b529-2ec461452791\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547457 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-utilities\") pod \"067264d3-8fc2-46cd-8d94-e89f9e25a500\" (UID: \"067264d3-8fc2-46cd-8d94-e89f9e25a500\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547491 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-utilities\") pod \"15d06ca2-19f5-4c88-8c32-51e491ffa834\" (UID: \"15d06ca2-19f5-4c88-8c32-51e491ffa834\") " Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547836 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547848 4692 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547858 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbw5p\" (UniqueName: \"kubernetes.io/projected/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-kube-api-access-sbw5p\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547868 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73819651-942b-4436-b0b1-7225cfb20ef1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547878 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pnq6\" (UniqueName: \"kubernetes.io/projected/15d06ca2-19f5-4c88-8c32-51e491ffa834-kube-api-access-2pnq6\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547887 4692 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/96013f1f-d2b0-42ec-a4f3-3626f48be8e1-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547895 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547903 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gkcn\" (UniqueName: \"kubernetes.io/projected/067264d3-8fc2-46cd-8d94-e89f9e25a500-kube-api-access-6gkcn\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.547911 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6gqq\" (UniqueName: \"kubernetes.io/projected/73819651-942b-4436-b0b1-7225cfb20ef1-kube-api-access-g6gqq\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.548799 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-utilities" (OuterVolumeSpecName: "utilities") pod "15d06ca2-19f5-4c88-8c32-51e491ffa834" (UID: "15d06ca2-19f5-4c88-8c32-51e491ffa834"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.550493 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d4b7447-a6e6-4530-b529-2ec461452791-kube-api-access-nzw78" (OuterVolumeSpecName: "kube-api-access-nzw78") pod "3d4b7447-a6e6-4530-b529-2ec461452791" (UID: "3d4b7447-a6e6-4530-b529-2ec461452791"). InnerVolumeSpecName "kube-api-access-nzw78". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.554792 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-utilities" (OuterVolumeSpecName: "utilities") pod "067264d3-8fc2-46cd-8d94-e89f9e25a500" (UID: "067264d3-8fc2-46cd-8d94-e89f9e25a500"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.557951 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d4b7447-a6e6-4530-b529-2ec461452791" (UID: "3d4b7447-a6e6-4530-b529-2ec461452791"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.614245 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "067264d3-8fc2-46cd-8d94-e89f9e25a500" (UID: "067264d3-8fc2-46cd-8d94-e89f9e25a500"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.648802 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzw78\" (UniqueName: \"kubernetes.io/projected/3d4b7447-a6e6-4530-b529-2ec461452791-kube-api-access-nzw78\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.648830 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.648841 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.648849 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/067264d3-8fc2-46cd-8d94-e89f9e25a500-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.648857 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d4b7447-a6e6-4530-b529-2ec461452791-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.685399 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "15d06ca2-19f5-4c88-8c32-51e491ffa834" (UID: "15d06ca2-19f5-4c88-8c32-51e491ffa834"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:25:54 crc kubenswrapper[4692]: I0309 09:25:54.751259 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15d06ca2-19f5-4c88-8c32-51e491ffa834-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.222718 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" event={"ID":"96013f1f-d2b0-42ec-a4f3-3626f48be8e1","Type":"ContainerDied","Data":"f16e39be91f32263c836c1e1974c4418ec7eb12ffb0525ae6c9188746542c632"} Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.222889 4692 scope.go:117] "RemoveContainer" containerID="4320322bc11ceae70df4eb69bbb20e01c609f6a78877297e606e084168324a84" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.223016 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdvhn" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.225411 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnf7m" event={"ID":"15d06ca2-19f5-4c88-8c32-51e491ffa834","Type":"ContainerDied","Data":"814b560c1411cc99b41a5e4bb6bb0713b89380ed4248d221aefeb3f3ce570860"} Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.225508 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnf7m" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.227225 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" event={"ID":"74fc14ae-20b4-429a-8f94-886a6bc85d3e","Type":"ContainerStarted","Data":"c32f94315c27ea3332843f9e7c4aa0decb40a63aa32a0772f4721e62e726f8cc"} Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.227257 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" event={"ID":"74fc14ae-20b4-429a-8f94-886a6bc85d3e","Type":"ContainerStarted","Data":"e28678d38d4dfb684caab379233d4063984d66828cd5681a2daf219fd3de79d4"} Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.227554 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.229444 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vkkhr" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.229634 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vkkhr" event={"ID":"3d4b7447-a6e6-4530-b529-2ec461452791","Type":"ContainerDied","Data":"71c712a5ba9cf979707ba7745db51f44a2867b60fb63282f872f99737e0e68c1"} Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.231005 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.238553 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4bqhb" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.239111 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ktmgp" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.240263 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ktmgp" event={"ID":"067264d3-8fc2-46cd-8d94-e89f9e25a500","Type":"ContainerDied","Data":"7b70955e661f01296c52dec402d93d90f896b537a065ed4a62baf31ed4e6fc31"} Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.245021 4692 scope.go:117] "RemoveContainer" containerID="2f11fd8acf0cbe68dc0911c7764a229e4c763847bce01b88fa085f3dd4ae6726" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.263834 4692 scope.go:117] "RemoveContainer" containerID="56620e11f32ad913c2c09495f85014c2118dd73419dcab9402f83f17cb3b28ce" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.327872 4692 scope.go:117] "RemoveContainer" containerID="104b76fb6f9d88f766ca6910007c7dc64151f713d824dfbc8ff9a211a91cb5ce" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.331469 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wvcsq" podStartSLOduration=2.331445189 podStartE2EDuration="2.331445189s" podCreationTimestamp="2026-03-09 09:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:25:55.259717953 +0000 UTC m=+356.084453534" watchObservedRunningTime="2026-03-09 09:25:55.331445189 +0000 UTC m=+356.156180770" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.353658 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdvhn"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.363775 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdvhn"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.387246 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pnf7m"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.403135 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pnf7m"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.403730 4692 scope.go:117] "RemoveContainer" containerID="cbde052a8bdb09ca7abc3e7e6fbe256b10a9973e4c03840951d8a6e9efac976a" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.412208 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4bqhb"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.417348 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4bqhb"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.422305 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkkhr"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.433547 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vkkhr"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.437479 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ktmgp"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.441405 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ktmgp"] Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.453652 4692 scope.go:117] "RemoveContainer" containerID="cca27779280f6a12e2d3ab6bbcc4fba1f3892575723d534aaea04333cc76bbd1" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.483002 4692 scope.go:117] "RemoveContainer" containerID="08664efca426d1881961484a5961fc666e14026ba62b4d2653ee4e236c61172c" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.498782 4692 scope.go:117] "RemoveContainer" containerID="efea09f529044bf3f09eca867fd99b9e990d01df4cf75b431b94df5261d9fb7b" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.516128 4692 scope.go:117] "RemoveContainer" containerID="7bc9bef4716ed588151c98669de90ff155904ec54de77f6e824da71c09e71750" Mar 09 09:25:55 crc kubenswrapper[4692]: I0309 09:25:55.534720 4692 scope.go:117] "RemoveContainer" containerID="3ed3288e4cde1e8118367b9fac99d39dcb77cb82dc3e097c1ce2566f60a7a2bd" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.084085 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" path="/var/lib/kubelet/pods/067264d3-8fc2-46cd-8d94-e89f9e25a500/volumes" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.084852 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" path="/var/lib/kubelet/pods/15d06ca2-19f5-4c88-8c32-51e491ffa834/volumes" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.085565 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" path="/var/lib/kubelet/pods/3d4b7447-a6e6-4530-b529-2ec461452791/volumes" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.088538 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" path="/var/lib/kubelet/pods/73819651-942b-4436-b0b1-7225cfb20ef1/volumes" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.089212 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" path="/var/lib/kubelet/pods/96013f1f-d2b0-42ec-a4f3-3626f48be8e1/volumes" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.095296 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gfw8w"] Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.095719 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.095829 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.095916 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.095986 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096059 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096129 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096245 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096260 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096270 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096278 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096289 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096298 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096308 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096315 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096324 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096330 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096341 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096348 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096369 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096377 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096389 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096396 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="extract-content" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096409 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096416 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096426 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096434 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="extract-utilities" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096548 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d4b7447-a6e6-4530-b529-2ec461452791" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096571 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096583 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="73819651-942b-4436-b0b1-7225cfb20ef1" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096595 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="067264d3-8fc2-46cd-8d94-e89f9e25a500" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096605 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="15d06ca2-19f5-4c88-8c32-51e491ffa834" containerName="registry-server" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096613 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" Mar 09 09:25:56 crc kubenswrapper[4692]: E0309 09:25:56.096723 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.096736 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="96013f1f-d2b0-42ec-a4f3-3626f48be8e1" containerName="marketplace-operator" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.097531 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.101020 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.101267 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gfw8w"] Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.276659 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8065edd5-e74a-439a-8e0a-e198d4915738-catalog-content\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.277053 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggqs7\" (UniqueName: \"kubernetes.io/projected/8065edd5-e74a-439a-8e0a-e198d4915738-kube-api-access-ggqs7\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.277214 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8065edd5-e74a-439a-8e0a-e198d4915738-utilities\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.379459 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggqs7\" (UniqueName: \"kubernetes.io/projected/8065edd5-e74a-439a-8e0a-e198d4915738-kube-api-access-ggqs7\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.379616 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8065edd5-e74a-439a-8e0a-e198d4915738-utilities\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.379652 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8065edd5-e74a-439a-8e0a-e198d4915738-catalog-content\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.380146 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8065edd5-e74a-439a-8e0a-e198d4915738-catalog-content\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.380716 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8065edd5-e74a-439a-8e0a-e198d4915738-utilities\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.402371 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggqs7\" (UniqueName: \"kubernetes.io/projected/8065edd5-e74a-439a-8e0a-e198d4915738-kube-api-access-ggqs7\") pod \"certified-operators-gfw8w\" (UID: \"8065edd5-e74a-439a-8e0a-e198d4915738\") " pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.420429 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:25:56 crc kubenswrapper[4692]: I0309 09:25:56.827506 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gfw8w"] Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.093544 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pjjn2"] Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.094776 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.096331 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.098086 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pjjn2"] Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.189931 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be7a257-3244-484e-9747-08d81edfd021-utilities\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.189982 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/2be7a257-3244-484e-9747-08d81edfd021-kube-api-access-r5lkh\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.190009 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be7a257-3244-484e-9747-08d81edfd021-catalog-content\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.252128 4692 generic.go:334] "Generic (PLEG): container finished" podID="8065edd5-e74a-439a-8e0a-e198d4915738" containerID="52d941aa812e7bde9a505bbb7d195b0d70c9190c3416cacde9433babb235361b" exitCode=0 Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.252199 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfw8w" event={"ID":"8065edd5-e74a-439a-8e0a-e198d4915738","Type":"ContainerDied","Data":"52d941aa812e7bde9a505bbb7d195b0d70c9190c3416cacde9433babb235361b"} Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.252284 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfw8w" event={"ID":"8065edd5-e74a-439a-8e0a-e198d4915738","Type":"ContainerStarted","Data":"22f917f946a3cc3e8e413a07c556936a1c93a86c321726f632ba9b15d71936a4"} Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.291823 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be7a257-3244-484e-9747-08d81edfd021-utilities\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.291873 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/2be7a257-3244-484e-9747-08d81edfd021-kube-api-access-r5lkh\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.292465 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be7a257-3244-484e-9747-08d81edfd021-catalog-content\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.292530 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2be7a257-3244-484e-9747-08d81edfd021-utilities\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.292794 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2be7a257-3244-484e-9747-08d81edfd021-catalog-content\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.313820 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5lkh\" (UniqueName: \"kubernetes.io/projected/2be7a257-3244-484e-9747-08d81edfd021-kube-api-access-r5lkh\") pod \"redhat-marketplace-pjjn2\" (UID: \"2be7a257-3244-484e-9747-08d81edfd021\") " pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.415417 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:25:57 crc kubenswrapper[4692]: I0309 09:25:57.808846 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pjjn2"] Mar 09 09:25:57 crc kubenswrapper[4692]: W0309 09:25:57.825587 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2be7a257_3244_484e_9747_08d81edfd021.slice/crio-1456689601fcd801fffa8c566983be27be79f527194da3ab701913dae4b9b1fa WatchSource:0}: Error finding container 1456689601fcd801fffa8c566983be27be79f527194da3ab701913dae4b9b1fa: Status 404 returned error can't find the container with id 1456689601fcd801fffa8c566983be27be79f527194da3ab701913dae4b9b1fa Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.258219 4692 generic.go:334] "Generic (PLEG): container finished" podID="2be7a257-3244-484e-9747-08d81edfd021" containerID="dff1e6f91fdf2fe50dd297d2ae64339a037169a7aea172184abc01b318457919" exitCode=0 Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.258683 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjjn2" event={"ID":"2be7a257-3244-484e-9747-08d81edfd021","Type":"ContainerDied","Data":"dff1e6f91fdf2fe50dd297d2ae64339a037169a7aea172184abc01b318457919"} Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.258712 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjjn2" event={"ID":"2be7a257-3244-484e-9747-08d81edfd021","Type":"ContainerStarted","Data":"1456689601fcd801fffa8c566983be27be79f527194da3ab701913dae4b9b1fa"} Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.262479 4692 generic.go:334] "Generic (PLEG): container finished" podID="8065edd5-e74a-439a-8e0a-e198d4915738" containerID="7c4de08f82204bd39d898edb1e71bea84c5e60a79926151c3e29495d5c4019db" exitCode=0 Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.262560 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfw8w" event={"ID":"8065edd5-e74a-439a-8e0a-e198d4915738","Type":"ContainerDied","Data":"7c4de08f82204bd39d898edb1e71bea84c5e60a79926151c3e29495d5c4019db"} Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.680952 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n6vw9"] Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.682249 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.684915 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.696429 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6vw9"] Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.716431 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmq2k\" (UniqueName: \"kubernetes.io/projected/335a9dc7-133c-4667-bb5f-3ba7598aeefc-kube-api-access-kmq2k\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.716489 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335a9dc7-133c-4667-bb5f-3ba7598aeefc-catalog-content\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.716544 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335a9dc7-133c-4667-bb5f-3ba7598aeefc-utilities\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.817973 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335a9dc7-133c-4667-bb5f-3ba7598aeefc-utilities\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.818053 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmq2k\" (UniqueName: \"kubernetes.io/projected/335a9dc7-133c-4667-bb5f-3ba7598aeefc-kube-api-access-kmq2k\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.818095 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335a9dc7-133c-4667-bb5f-3ba7598aeefc-catalog-content\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.818538 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/335a9dc7-133c-4667-bb5f-3ba7598aeefc-catalog-content\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.819109 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/335a9dc7-133c-4667-bb5f-3ba7598aeefc-utilities\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:58 crc kubenswrapper[4692]: I0309 09:25:58.841267 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmq2k\" (UniqueName: \"kubernetes.io/projected/335a9dc7-133c-4667-bb5f-3ba7598aeefc-kube-api-access-kmq2k\") pod \"redhat-operators-n6vw9\" (UID: \"335a9dc7-133c-4667-bb5f-3ba7598aeefc\") " pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.037301 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.276953 4692 generic.go:334] "Generic (PLEG): container finished" podID="2be7a257-3244-484e-9747-08d81edfd021" containerID="a8c79dd66ed615acf1e1e0e3dc3610f2bc9e136f1b811c17de2121f94ec8c28a" exitCode=0 Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.277191 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjjn2" event={"ID":"2be7a257-3244-484e-9747-08d81edfd021","Type":"ContainerDied","Data":"a8c79dd66ed615acf1e1e0e3dc3610f2bc9e136f1b811c17de2121f94ec8c28a"} Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.282663 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gfw8w" event={"ID":"8065edd5-e74a-439a-8e0a-e198d4915738","Type":"ContainerStarted","Data":"eb43592cedba7ebe7afe3af5bd041c7aaff018a910f6653dd4e2aa4b86807efe"} Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.328090 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gfw8w" podStartSLOduration=1.8005784390000001 podStartE2EDuration="3.328071572s" podCreationTimestamp="2026-03-09 09:25:56 +0000 UTC" firstStartedPulling="2026-03-09 09:25:57.253547712 +0000 UTC m=+358.078283293" lastFinishedPulling="2026-03-09 09:25:58.781040845 +0000 UTC m=+359.605776426" observedRunningTime="2026-03-09 09:25:59.320317071 +0000 UTC m=+360.145052662" watchObservedRunningTime="2026-03-09 09:25:59.328071572 +0000 UTC m=+360.152807153" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.480280 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n6vw9"] Mar 09 09:25:59 crc kubenswrapper[4692]: W0309 09:25:59.496763 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod335a9dc7_133c_4667_bb5f_3ba7598aeefc.slice/crio-0413251e162289c4769d75b0dd1e8bc6205d1088758f922b24d57cea74fc68c4 WatchSource:0}: Error finding container 0413251e162289c4769d75b0dd1e8bc6205d1088758f922b24d57cea74fc68c4: Status 404 returned error can't find the container with id 0413251e162289c4769d75b0dd1e8bc6205d1088758f922b24d57cea74fc68c4 Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.684145 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-68gtj"] Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.685641 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.687903 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.696904 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68gtj"] Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.832028 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n4b5\" (UniqueName: \"kubernetes.io/projected/1df3c8b1-25ba-4b83-abdd-bf70af81208c-kube-api-access-4n4b5\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.832088 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-utilities\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.832263 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-catalog-content\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.933860 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n4b5\" (UniqueName: \"kubernetes.io/projected/1df3c8b1-25ba-4b83-abdd-bf70af81208c-kube-api-access-4n4b5\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.933925 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-utilities\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.933947 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-catalog-content\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.934437 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-catalog-content\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.934539 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-utilities\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:25:59 crc kubenswrapper[4692]: I0309 09:25:59.963427 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n4b5\" (UniqueName: \"kubernetes.io/projected/1df3c8b1-25ba-4b83-abdd-bf70af81208c-kube-api-access-4n4b5\") pod \"community-operators-68gtj\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.091900 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.098666 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.141049 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550806-99ld4"] Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.144756 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550806-99ld4" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.147656 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.148076 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.148329 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.148687 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550806-99ld4"] Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.294605 4692 generic.go:334] "Generic (PLEG): container finished" podID="335a9dc7-133c-4667-bb5f-3ba7598aeefc" containerID="56a2cd482a212bc0384a9f6177a08db12f2b68075c10582369efdd71f39a4401" exitCode=0 Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.295812 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6vw9" event={"ID":"335a9dc7-133c-4667-bb5f-3ba7598aeefc","Type":"ContainerDied","Data":"56a2cd482a212bc0384a9f6177a08db12f2b68075c10582369efdd71f39a4401"} Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.295854 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6vw9" event={"ID":"335a9dc7-133c-4667-bb5f-3ba7598aeefc","Type":"ContainerStarted","Data":"0413251e162289c4769d75b0dd1e8bc6205d1088758f922b24d57cea74fc68c4"} Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.303753 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjjn2" event={"ID":"2be7a257-3244-484e-9747-08d81edfd021","Type":"ContainerStarted","Data":"778fb0dbb5b051625e3c6d3470bc49620cf2433a794aa1ec1c6118448c8d4f2a"} Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.340857 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qjbs\" (UniqueName: \"kubernetes.io/projected/9f98f499-bbc9-4ae7-b470-dfe2a771127e-kube-api-access-8qjbs\") pod \"auto-csr-approver-29550806-99ld4\" (UID: \"9f98f499-bbc9-4ae7-b470-dfe2a771127e\") " pod="openshift-infra/auto-csr-approver-29550806-99ld4" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.344941 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pjjn2" podStartSLOduration=1.966115141 podStartE2EDuration="3.3449199s" podCreationTimestamp="2026-03-09 09:25:57 +0000 UTC" firstStartedPulling="2026-03-09 09:25:58.260953293 +0000 UTC m=+359.085688874" lastFinishedPulling="2026-03-09 09:25:59.639758052 +0000 UTC m=+360.464493633" observedRunningTime="2026-03-09 09:26:00.336520371 +0000 UTC m=+361.161255952" watchObservedRunningTime="2026-03-09 09:26:00.3449199 +0000 UTC m=+361.169655471" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.441667 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qjbs\" (UniqueName: \"kubernetes.io/projected/9f98f499-bbc9-4ae7-b470-dfe2a771127e-kube-api-access-8qjbs\") pod \"auto-csr-approver-29550806-99ld4\" (UID: \"9f98f499-bbc9-4ae7-b470-dfe2a771127e\") " pod="openshift-infra/auto-csr-approver-29550806-99ld4" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.463837 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qjbs\" (UniqueName: \"kubernetes.io/projected/9f98f499-bbc9-4ae7-b470-dfe2a771127e-kube-api-access-8qjbs\") pod \"auto-csr-approver-29550806-99ld4\" (UID: \"9f98f499-bbc9-4ae7-b470-dfe2a771127e\") " pod="openshift-infra/auto-csr-approver-29550806-99ld4" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.473621 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550806-99ld4" Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.542471 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68gtj"] Mar 09 09:26:00 crc kubenswrapper[4692]: W0309 09:26:00.554345 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1df3c8b1_25ba_4b83_abdd_bf70af81208c.slice/crio-30aa005897999dcd1f7e231530f8409996567fc2b305ec9ae23cbda04825498d WatchSource:0}: Error finding container 30aa005897999dcd1f7e231530f8409996567fc2b305ec9ae23cbda04825498d: Status 404 returned error can't find the container with id 30aa005897999dcd1f7e231530f8409996567fc2b305ec9ae23cbda04825498d Mar 09 09:26:00 crc kubenswrapper[4692]: I0309 09:26:00.918907 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550806-99ld4"] Mar 09 09:26:00 crc kubenswrapper[4692]: W0309 09:26:00.924524 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f98f499_bbc9_4ae7_b470_dfe2a771127e.slice/crio-65702520f001bd97aa3abd906587d89ec4f29a9df99343a151bbf7a3282a03c8 WatchSource:0}: Error finding container 65702520f001bd97aa3abd906587d89ec4f29a9df99343a151bbf7a3282a03c8: Status 404 returned error can't find the container with id 65702520f001bd97aa3abd906587d89ec4f29a9df99343a151bbf7a3282a03c8 Mar 09 09:26:01 crc kubenswrapper[4692]: I0309 09:26:01.310536 4692 generic.go:334] "Generic (PLEG): container finished" podID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerID="c2217d668c24d290125792a77e8944b4ebee1c33814afbb22098ae14a53819fe" exitCode=0 Mar 09 09:26:01 crc kubenswrapper[4692]: I0309 09:26:01.310668 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68gtj" event={"ID":"1df3c8b1-25ba-4b83-abdd-bf70af81208c","Type":"ContainerDied","Data":"c2217d668c24d290125792a77e8944b4ebee1c33814afbb22098ae14a53819fe"} Mar 09 09:26:01 crc kubenswrapper[4692]: I0309 09:26:01.310701 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68gtj" event={"ID":"1df3c8b1-25ba-4b83-abdd-bf70af81208c","Type":"ContainerStarted","Data":"30aa005897999dcd1f7e231530f8409996567fc2b305ec9ae23cbda04825498d"} Mar 09 09:26:01 crc kubenswrapper[4692]: I0309 09:26:01.315009 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550806-99ld4" event={"ID":"9f98f499-bbc9-4ae7-b470-dfe2a771127e","Type":"ContainerStarted","Data":"65702520f001bd97aa3abd906587d89ec4f29a9df99343a151bbf7a3282a03c8"} Mar 09 09:26:02 crc kubenswrapper[4692]: I0309 09:26:02.323755 4692 generic.go:334] "Generic (PLEG): container finished" podID="335a9dc7-133c-4667-bb5f-3ba7598aeefc" containerID="ac7c7850032c11e5723273c00068848773a550315cf3af9f032ecfce369dde35" exitCode=0 Mar 09 09:26:02 crc kubenswrapper[4692]: I0309 09:26:02.323817 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6vw9" event={"ID":"335a9dc7-133c-4667-bb5f-3ba7598aeefc","Type":"ContainerDied","Data":"ac7c7850032c11e5723273c00068848773a550315cf3af9f032ecfce369dde35"} Mar 09 09:26:02 crc kubenswrapper[4692]: I0309 09:26:02.328103 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550806-99ld4" event={"ID":"9f98f499-bbc9-4ae7-b470-dfe2a771127e","Type":"ContainerStarted","Data":"162b6ea9823b8b3b412a77b68ee22dcc0eb9b5dcf4a5abf93fd06d9d6832a578"} Mar 09 09:26:02 crc kubenswrapper[4692]: I0309 09:26:02.336425 4692 generic.go:334] "Generic (PLEG): container finished" podID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerID="d575fc5324e973ef932620dae0ac1c8d6d04170c3dcc03813ebfd838ac1618b4" exitCode=0 Mar 09 09:26:02 crc kubenswrapper[4692]: I0309 09:26:02.336490 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68gtj" event={"ID":"1df3c8b1-25ba-4b83-abdd-bf70af81208c","Type":"ContainerDied","Data":"d575fc5324e973ef932620dae0ac1c8d6d04170c3dcc03813ebfd838ac1618b4"} Mar 09 09:26:02 crc kubenswrapper[4692]: I0309 09:26:02.400089 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29550806-99ld4" podStartSLOduration=1.455882374 podStartE2EDuration="2.40004746s" podCreationTimestamp="2026-03-09 09:26:00 +0000 UTC" firstStartedPulling="2026-03-09 09:26:00.926437477 +0000 UTC m=+361.751173058" lastFinishedPulling="2026-03-09 09:26:01.870602563 +0000 UTC m=+362.695338144" observedRunningTime="2026-03-09 09:26:02.391788455 +0000 UTC m=+363.216524056" watchObservedRunningTime="2026-03-09 09:26:02.40004746 +0000 UTC m=+363.224783061" Mar 09 09:26:03 crc kubenswrapper[4692]: I0309 09:26:03.374033 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n6vw9" event={"ID":"335a9dc7-133c-4667-bb5f-3ba7598aeefc","Type":"ContainerStarted","Data":"5071f9329db3e9b6d89882f62dd3d8272a530664fcc785a908b67256825c60f0"} Mar 09 09:26:03 crc kubenswrapper[4692]: I0309 09:26:03.376592 4692 generic.go:334] "Generic (PLEG): container finished" podID="9f98f499-bbc9-4ae7-b470-dfe2a771127e" containerID="162b6ea9823b8b3b412a77b68ee22dcc0eb9b5dcf4a5abf93fd06d9d6832a578" exitCode=0 Mar 09 09:26:03 crc kubenswrapper[4692]: I0309 09:26:03.376660 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550806-99ld4" event={"ID":"9f98f499-bbc9-4ae7-b470-dfe2a771127e","Type":"ContainerDied","Data":"162b6ea9823b8b3b412a77b68ee22dcc0eb9b5dcf4a5abf93fd06d9d6832a578"} Mar 09 09:26:03 crc kubenswrapper[4692]: I0309 09:26:03.380682 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68gtj" event={"ID":"1df3c8b1-25ba-4b83-abdd-bf70af81208c","Type":"ContainerStarted","Data":"522041d8e3397f7b3009880124ed49c62f56b5c13c5e15a8881aa7b83a50b75e"} Mar 09 09:26:03 crc kubenswrapper[4692]: I0309 09:26:03.396384 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n6vw9" podStartSLOduration=2.956145207 podStartE2EDuration="5.396364369s" podCreationTimestamp="2026-03-09 09:25:58 +0000 UTC" firstStartedPulling="2026-03-09 09:26:00.297783284 +0000 UTC m=+361.122518865" lastFinishedPulling="2026-03-09 09:26:02.738002446 +0000 UTC m=+363.562738027" observedRunningTime="2026-03-09 09:26:03.395505156 +0000 UTC m=+364.220240757" watchObservedRunningTime="2026-03-09 09:26:03.396364369 +0000 UTC m=+364.221099950" Mar 09 09:26:03 crc kubenswrapper[4692]: I0309 09:26:03.415939 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-68gtj" podStartSLOduration=2.95698195 podStartE2EDuration="4.415920533s" podCreationTimestamp="2026-03-09 09:25:59 +0000 UTC" firstStartedPulling="2026-03-09 09:26:01.314349785 +0000 UTC m=+362.139085366" lastFinishedPulling="2026-03-09 09:26:02.773288328 +0000 UTC m=+363.598023949" observedRunningTime="2026-03-09 09:26:03.415535162 +0000 UTC m=+364.240270743" watchObservedRunningTime="2026-03-09 09:26:03.415920533 +0000 UTC m=+364.240656114" Mar 09 09:26:04 crc kubenswrapper[4692]: I0309 09:26:04.637775 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550806-99ld4" Mar 09 09:26:04 crc kubenswrapper[4692]: I0309 09:26:04.810306 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qjbs\" (UniqueName: \"kubernetes.io/projected/9f98f499-bbc9-4ae7-b470-dfe2a771127e-kube-api-access-8qjbs\") pod \"9f98f499-bbc9-4ae7-b470-dfe2a771127e\" (UID: \"9f98f499-bbc9-4ae7-b470-dfe2a771127e\") " Mar 09 09:26:04 crc kubenswrapper[4692]: I0309 09:26:04.826376 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f98f499-bbc9-4ae7-b470-dfe2a771127e-kube-api-access-8qjbs" (OuterVolumeSpecName: "kube-api-access-8qjbs") pod "9f98f499-bbc9-4ae7-b470-dfe2a771127e" (UID: "9f98f499-bbc9-4ae7-b470-dfe2a771127e"). InnerVolumeSpecName "kube-api-access-8qjbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:26:04 crc kubenswrapper[4692]: I0309 09:26:04.911916 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qjbs\" (UniqueName: \"kubernetes.io/projected/9f98f499-bbc9-4ae7-b470-dfe2a771127e-kube-api-access-8qjbs\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:05 crc kubenswrapper[4692]: I0309 09:26:05.391768 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550806-99ld4" event={"ID":"9f98f499-bbc9-4ae7-b470-dfe2a771127e","Type":"ContainerDied","Data":"65702520f001bd97aa3abd906587d89ec4f29a9df99343a151bbf7a3282a03c8"} Mar 09 09:26:05 crc kubenswrapper[4692]: I0309 09:26:05.391818 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65702520f001bd97aa3abd906587d89ec4f29a9df99343a151bbf7a3282a03c8" Mar 09 09:26:05 crc kubenswrapper[4692]: I0309 09:26:05.392143 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550806-99ld4" Mar 09 09:26:06 crc kubenswrapper[4692]: I0309 09:26:06.421477 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:26:06 crc kubenswrapper[4692]: I0309 09:26:06.421586 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:26:06 crc kubenswrapper[4692]: I0309 09:26:06.464570 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:26:07 crc kubenswrapper[4692]: I0309 09:26:07.416008 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:26:07 crc kubenswrapper[4692]: I0309 09:26:07.416068 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:26:07 crc kubenswrapper[4692]: I0309 09:26:07.449150 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:26:07 crc kubenswrapper[4692]: I0309 09:26:07.452106 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gfw8w" Mar 09 09:26:08 crc kubenswrapper[4692]: I0309 09:26:08.453707 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pjjn2" Mar 09 09:26:09 crc kubenswrapper[4692]: I0309 09:26:09.037942 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:26:09 crc kubenswrapper[4692]: I0309 09:26:09.038003 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:26:09 crc kubenswrapper[4692]: I0309 09:26:09.084921 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:26:09 crc kubenswrapper[4692]: I0309 09:26:09.461792 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n6vw9" Mar 09 09:26:10 crc kubenswrapper[4692]: I0309 09:26:10.099650 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:26:10 crc kubenswrapper[4692]: I0309 09:26:10.099704 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:26:10 crc kubenswrapper[4692]: I0309 09:26:10.145276 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:26:10 crc kubenswrapper[4692]: I0309 09:26:10.456092 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.336402 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" podUID="946e899e-f89e-4f66-80a2-1f3effc06b7f" containerName="registry" containerID="cri-o://9d8f07d1dd0caa34ab1801883390d68844fd2f19b56dee0acca723bd3cd9823d" gracePeriod=30 Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.458000 4692 generic.go:334] "Generic (PLEG): container finished" podID="946e899e-f89e-4f66-80a2-1f3effc06b7f" containerID="9d8f07d1dd0caa34ab1801883390d68844fd2f19b56dee0acca723bd3cd9823d" exitCode=0 Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.458126 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" event={"ID":"946e899e-f89e-4f66-80a2-1f3effc06b7f","Type":"ContainerDied","Data":"9d8f07d1dd0caa34ab1801883390d68844fd2f19b56dee0acca723bd3cd9823d"} Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.701413 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778321 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-bound-sa-token\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778375 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946e899e-f89e-4f66-80a2-1f3effc06b7f-ca-trust-extracted\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778396 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-tls\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778435 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t88ck\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-kube-api-access-t88ck\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778459 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-trusted-ca\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778500 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946e899e-f89e-4f66-80a2-1f3effc06b7f-installation-pull-secrets\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778519 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-certificates\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.778705 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"946e899e-f89e-4f66-80a2-1f3effc06b7f\" (UID: \"946e899e-f89e-4f66-80a2-1f3effc06b7f\") " Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.779656 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.779738 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.784516 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/946e899e-f89e-4f66-80a2-1f3effc06b7f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.785016 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-kube-api-access-t88ck" (OuterVolumeSpecName: "kube-api-access-t88ck") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "kube-api-access-t88ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.786661 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.787478 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.790768 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.794570 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/946e899e-f89e-4f66-80a2-1f3effc06b7f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "946e899e-f89e-4f66-80a2-1f3effc06b7f" (UID: "946e899e-f89e-4f66-80a2-1f3effc06b7f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.880566 4692 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/946e899e-f89e-4f66-80a2-1f3effc06b7f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.880606 4692 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.880615 4692 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.880626 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t88ck\" (UniqueName: \"kubernetes.io/projected/946e899e-f89e-4f66-80a2-1f3effc06b7f-kube-api-access-t88ck\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.880636 4692 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.880645 4692 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/946e899e-f89e-4f66-80a2-1f3effc06b7f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:18 crc kubenswrapper[4692]: I0309 09:26:18.880654 4692 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/946e899e-f89e-4f66-80a2-1f3effc06b7f-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 09 09:26:19 crc kubenswrapper[4692]: I0309 09:26:19.465677 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" event={"ID":"946e899e-f89e-4f66-80a2-1f3effc06b7f","Type":"ContainerDied","Data":"2302f75aacff21fe204ca67e0ed8bc3dd242d7e995da02f4181eec23548af690"} Mar 09 09:26:19 crc kubenswrapper[4692]: I0309 09:26:19.465997 4692 scope.go:117] "RemoveContainer" containerID="9d8f07d1dd0caa34ab1801883390d68844fd2f19b56dee0acca723bd3cd9823d" Mar 09 09:26:19 crc kubenswrapper[4692]: I0309 09:26:19.465763 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxwb2" Mar 09 09:26:19 crc kubenswrapper[4692]: I0309 09:26:19.510972 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxwb2"] Mar 09 09:26:19 crc kubenswrapper[4692]: I0309 09:26:19.517786 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxwb2"] Mar 09 09:26:20 crc kubenswrapper[4692]: I0309 09:26:20.079134 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946e899e-f89e-4f66-80a2-1f3effc06b7f" path="/var/lib/kubelet/pods/946e899e-f89e-4f66-80a2-1f3effc06b7f/volumes" Mar 09 09:26:47 crc kubenswrapper[4692]: I0309 09:26:47.624033 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:26:47 crc kubenswrapper[4692]: I0309 09:26:47.624504 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:27:17 crc kubenswrapper[4692]: I0309 09:27:17.624224 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:27:17 crc kubenswrapper[4692]: I0309 09:27:17.624779 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.624234 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.624749 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.624785 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.625292 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61fae0ebf34e8ab848fb0659a787a8b14e18310fadde50c12777c1d7eafc19fe"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.625347 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://61fae0ebf34e8ab848fb0659a787a8b14e18310fadde50c12777c1d7eafc19fe" gracePeriod=600 Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.920306 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="61fae0ebf34e8ab848fb0659a787a8b14e18310fadde50c12777c1d7eafc19fe" exitCode=0 Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.920390 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"61fae0ebf34e8ab848fb0659a787a8b14e18310fadde50c12777c1d7eafc19fe"} Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.920642 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"31f9b48232438cf31701990fc67d0a37a49524e82d3bf2636f02423390a87487"} Mar 09 09:27:47 crc kubenswrapper[4692]: I0309 09:27:47.920660 4692 scope.go:117] "RemoveContainer" containerID="b30ec1f14935ee72de8d696ab21516444d1dbed8e188660550c2b615537f1de4" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.133062 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550808-shsh8"] Mar 09 09:28:00 crc kubenswrapper[4692]: E0309 09:28:00.133836 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946e899e-f89e-4f66-80a2-1f3effc06b7f" containerName="registry" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.133851 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="946e899e-f89e-4f66-80a2-1f3effc06b7f" containerName="registry" Mar 09 09:28:00 crc kubenswrapper[4692]: E0309 09:28:00.133871 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f98f499-bbc9-4ae7-b470-dfe2a771127e" containerName="oc" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.133880 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f98f499-bbc9-4ae7-b470-dfe2a771127e" containerName="oc" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.134002 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="946e899e-f89e-4f66-80a2-1f3effc06b7f" containerName="registry" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.134016 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f98f499-bbc9-4ae7-b470-dfe2a771127e" containerName="oc" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.134458 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550808-shsh8" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.137414 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.137607 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.137783 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.142193 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550808-shsh8"] Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.315575 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnt4g\" (UniqueName: \"kubernetes.io/projected/196f410f-b0a7-4bac-851f-2910d63c5b0f-kube-api-access-nnt4g\") pod \"auto-csr-approver-29550808-shsh8\" (UID: \"196f410f-b0a7-4bac-851f-2910d63c5b0f\") " pod="openshift-infra/auto-csr-approver-29550808-shsh8" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.417643 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnt4g\" (UniqueName: \"kubernetes.io/projected/196f410f-b0a7-4bac-851f-2910d63c5b0f-kube-api-access-nnt4g\") pod \"auto-csr-approver-29550808-shsh8\" (UID: \"196f410f-b0a7-4bac-851f-2910d63c5b0f\") " pod="openshift-infra/auto-csr-approver-29550808-shsh8" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.435691 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnt4g\" (UniqueName: \"kubernetes.io/projected/196f410f-b0a7-4bac-851f-2910d63c5b0f-kube-api-access-nnt4g\") pod \"auto-csr-approver-29550808-shsh8\" (UID: \"196f410f-b0a7-4bac-851f-2910d63c5b0f\") " pod="openshift-infra/auto-csr-approver-29550808-shsh8" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.476069 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550808-shsh8" Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.642289 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550808-shsh8"] Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.654267 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 09:28:00 crc kubenswrapper[4692]: I0309 09:28:00.990649 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550808-shsh8" event={"ID":"196f410f-b0a7-4bac-851f-2910d63c5b0f","Type":"ContainerStarted","Data":"618c493fcf122905250c80b17a25b0988ae13067b039d105298a5cae76c09dbd"} Mar 09 09:28:01 crc kubenswrapper[4692]: I0309 09:28:01.996632 4692 generic.go:334] "Generic (PLEG): container finished" podID="196f410f-b0a7-4bac-851f-2910d63c5b0f" containerID="530727072ce5566044dbb1c8874daac2dc304d4e46a651b7cf913f6748ed3967" exitCode=0 Mar 09 09:28:01 crc kubenswrapper[4692]: I0309 09:28:01.996675 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550808-shsh8" event={"ID":"196f410f-b0a7-4bac-851f-2910d63c5b0f","Type":"ContainerDied","Data":"530727072ce5566044dbb1c8874daac2dc304d4e46a651b7cf913f6748ed3967"} Mar 09 09:28:03 crc kubenswrapper[4692]: I0309 09:28:03.181804 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550808-shsh8" Mar 09 09:28:03 crc kubenswrapper[4692]: I0309 09:28:03.354139 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnt4g\" (UniqueName: \"kubernetes.io/projected/196f410f-b0a7-4bac-851f-2910d63c5b0f-kube-api-access-nnt4g\") pod \"196f410f-b0a7-4bac-851f-2910d63c5b0f\" (UID: \"196f410f-b0a7-4bac-851f-2910d63c5b0f\") " Mar 09 09:28:03 crc kubenswrapper[4692]: I0309 09:28:03.363049 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/196f410f-b0a7-4bac-851f-2910d63c5b0f-kube-api-access-nnt4g" (OuterVolumeSpecName: "kube-api-access-nnt4g") pod "196f410f-b0a7-4bac-851f-2910d63c5b0f" (UID: "196f410f-b0a7-4bac-851f-2910d63c5b0f"). InnerVolumeSpecName "kube-api-access-nnt4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:28:03 crc kubenswrapper[4692]: I0309 09:28:03.456382 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnt4g\" (UniqueName: \"kubernetes.io/projected/196f410f-b0a7-4bac-851f-2910d63c5b0f-kube-api-access-nnt4g\") on node \"crc\" DevicePath \"\"" Mar 09 09:28:04 crc kubenswrapper[4692]: I0309 09:28:04.010776 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550808-shsh8" event={"ID":"196f410f-b0a7-4bac-851f-2910d63c5b0f","Type":"ContainerDied","Data":"618c493fcf122905250c80b17a25b0988ae13067b039d105298a5cae76c09dbd"} Mar 09 09:28:04 crc kubenswrapper[4692]: I0309 09:28:04.010822 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="618c493fcf122905250c80b17a25b0988ae13067b039d105298a5cae76c09dbd" Mar 09 09:28:04 crc kubenswrapper[4692]: I0309 09:28:04.010841 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550808-shsh8" Mar 09 09:28:04 crc kubenswrapper[4692]: I0309 09:28:04.234098 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550802-w82g5"] Mar 09 09:28:04 crc kubenswrapper[4692]: I0309 09:28:04.238811 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550802-w82g5"] Mar 09 09:28:06 crc kubenswrapper[4692]: I0309 09:28:06.077147 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c0a3f4-29f0-42fa-83bc-83e15171ed77" path="/var/lib/kubelet/pods/64c0a3f4-29f0-42fa-83bc-83e15171ed77/volumes" Mar 09 09:29:00 crc kubenswrapper[4692]: I0309 09:29:00.778288 4692 scope.go:117] "RemoveContainer" containerID="57f97e200a45c7ffe5ff133937b61b7a6a5d987f6f2c7bdf4d137d62ad873142" Mar 09 09:29:00 crc kubenswrapper[4692]: I0309 09:29:00.793992 4692 scope.go:117] "RemoveContainer" containerID="b861a5a13ae14d622c134d6b2ea01cde37bbfae9efc6df4d37057cf71b528c86" Mar 09 09:29:00 crc kubenswrapper[4692]: I0309 09:29:00.827952 4692 scope.go:117] "RemoveContainer" containerID="77afccaf2045943b7772baffcde81fef46c22cefd5b6c5cbd70acc2a3c840789" Mar 09 09:29:47 crc kubenswrapper[4692]: I0309 09:29:47.623854 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:29:47 crc kubenswrapper[4692]: I0309 09:29:47.625322 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.134687 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550810-4ckmx"] Mar 09 09:30:00 crc kubenswrapper[4692]: E0309 09:30:00.135475 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196f410f-b0a7-4bac-851f-2910d63c5b0f" containerName="oc" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.135490 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="196f410f-b0a7-4bac-851f-2910d63c5b0f" containerName="oc" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.135605 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="196f410f-b0a7-4bac-851f-2910d63c5b0f" containerName="oc" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.136059 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550810-4ckmx" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.139538 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.139716 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.139840 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.147099 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550810-4ckmx"] Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.153824 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr"] Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.154696 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.157091 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.157408 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.167918 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr"] Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.303579 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzpwx\" (UniqueName: \"kubernetes.io/projected/69976e47-d9ef-4c41-8060-eb950d844750-kube-api-access-fzpwx\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.303649 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69976e47-d9ef-4c41-8060-eb950d844750-config-volume\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.304200 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69976e47-d9ef-4c41-8060-eb950d844750-secret-volume\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.304361 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76qw8\" (UniqueName: \"kubernetes.io/projected/4acdeb1a-b335-45c9-9ba6-22fe29675aaf-kube-api-access-76qw8\") pod \"auto-csr-approver-29550810-4ckmx\" (UID: \"4acdeb1a-b335-45c9-9ba6-22fe29675aaf\") " pod="openshift-infra/auto-csr-approver-29550810-4ckmx" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.405209 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzpwx\" (UniqueName: \"kubernetes.io/projected/69976e47-d9ef-4c41-8060-eb950d844750-kube-api-access-fzpwx\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.405258 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69976e47-d9ef-4c41-8060-eb950d844750-config-volume\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.405313 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69976e47-d9ef-4c41-8060-eb950d844750-secret-volume\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.405340 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76qw8\" (UniqueName: \"kubernetes.io/projected/4acdeb1a-b335-45c9-9ba6-22fe29675aaf-kube-api-access-76qw8\") pod \"auto-csr-approver-29550810-4ckmx\" (UID: \"4acdeb1a-b335-45c9-9ba6-22fe29675aaf\") " pod="openshift-infra/auto-csr-approver-29550810-4ckmx" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.406410 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69976e47-d9ef-4c41-8060-eb950d844750-config-volume\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.414801 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69976e47-d9ef-4c41-8060-eb950d844750-secret-volume\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.422122 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76qw8\" (UniqueName: \"kubernetes.io/projected/4acdeb1a-b335-45c9-9ba6-22fe29675aaf-kube-api-access-76qw8\") pod \"auto-csr-approver-29550810-4ckmx\" (UID: \"4acdeb1a-b335-45c9-9ba6-22fe29675aaf\") " pod="openshift-infra/auto-csr-approver-29550810-4ckmx" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.423002 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzpwx\" (UniqueName: \"kubernetes.io/projected/69976e47-d9ef-4c41-8060-eb950d844750-kube-api-access-fzpwx\") pod \"collect-profiles-29550810-4d2pr\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.458003 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550810-4ckmx" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.468741 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.639383 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550810-4ckmx"] Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.678703 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr"] Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.690620 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" event={"ID":"69976e47-d9ef-4c41-8060-eb950d844750","Type":"ContainerStarted","Data":"68045f2c90cda241041f6548df8d751ced3c5bfc7c5d4f8cf4b6afdb02d2bb7e"} Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.692026 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550810-4ckmx" event={"ID":"4acdeb1a-b335-45c9-9ba6-22fe29675aaf","Type":"ContainerStarted","Data":"c34fadce9dcdc4e91a3c0c85f50d786350dbade181b8e9ace93dee79b702b598"} Mar 09 09:30:00 crc kubenswrapper[4692]: I0309 09:30:00.866252 4692 scope.go:117] "RemoveContainer" containerID="6ec1ecbb6a65d8760942f519a7440b6a4dcfb4e6184b9b199bc82504b9635683" Mar 09 09:30:01 crc kubenswrapper[4692]: I0309 09:30:01.698932 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" event={"ID":"69976e47-d9ef-4c41-8060-eb950d844750","Type":"ContainerDied","Data":"5586f6d4d6c38d91a1aab1cc18dd0ba03fe78f3dbc08bfd2bc8679b5c3c6d1b8"} Mar 09 09:30:01 crc kubenswrapper[4692]: I0309 09:30:01.698738 4692 generic.go:334] "Generic (PLEG): container finished" podID="69976e47-d9ef-4c41-8060-eb950d844750" containerID="5586f6d4d6c38d91a1aab1cc18dd0ba03fe78f3dbc08bfd2bc8679b5c3c6d1b8" exitCode=0 Mar 09 09:30:02 crc kubenswrapper[4692]: I0309 09:30:02.706061 4692 generic.go:334] "Generic (PLEG): container finished" podID="4acdeb1a-b335-45c9-9ba6-22fe29675aaf" containerID="00a653d5284093aa21c801feafebffd170508a7c8231e5b02ee93ae4526a2393" exitCode=0 Mar 09 09:30:02 crc kubenswrapper[4692]: I0309 09:30:02.706144 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550810-4ckmx" event={"ID":"4acdeb1a-b335-45c9-9ba6-22fe29675aaf","Type":"ContainerDied","Data":"00a653d5284093aa21c801feafebffd170508a7c8231e5b02ee93ae4526a2393"} Mar 09 09:30:02 crc kubenswrapper[4692]: I0309 09:30:02.897895 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.040650 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69976e47-d9ef-4c41-8060-eb950d844750-config-volume\") pod \"69976e47-d9ef-4c41-8060-eb950d844750\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.040769 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69976e47-d9ef-4c41-8060-eb950d844750-secret-volume\") pod \"69976e47-d9ef-4c41-8060-eb950d844750\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.040791 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzpwx\" (UniqueName: \"kubernetes.io/projected/69976e47-d9ef-4c41-8060-eb950d844750-kube-api-access-fzpwx\") pod \"69976e47-d9ef-4c41-8060-eb950d844750\" (UID: \"69976e47-d9ef-4c41-8060-eb950d844750\") " Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.041597 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69976e47-d9ef-4c41-8060-eb950d844750-config-volume" (OuterVolumeSpecName: "config-volume") pod "69976e47-d9ef-4c41-8060-eb950d844750" (UID: "69976e47-d9ef-4c41-8060-eb950d844750"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.046385 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69976e47-d9ef-4c41-8060-eb950d844750-kube-api-access-fzpwx" (OuterVolumeSpecName: "kube-api-access-fzpwx") pod "69976e47-d9ef-4c41-8060-eb950d844750" (UID: "69976e47-d9ef-4c41-8060-eb950d844750"). InnerVolumeSpecName "kube-api-access-fzpwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.046587 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69976e47-d9ef-4c41-8060-eb950d844750-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "69976e47-d9ef-4c41-8060-eb950d844750" (UID: "69976e47-d9ef-4c41-8060-eb950d844750"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.142622 4692 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/69976e47-d9ef-4c41-8060-eb950d844750-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.142647 4692 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/69976e47-d9ef-4c41-8060-eb950d844750-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.142658 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzpwx\" (UniqueName: \"kubernetes.io/projected/69976e47-d9ef-4c41-8060-eb950d844750-kube-api-access-fzpwx\") on node \"crc\" DevicePath \"\"" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.712989 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.712989 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550810-4d2pr" event={"ID":"69976e47-d9ef-4c41-8060-eb950d844750","Type":"ContainerDied","Data":"68045f2c90cda241041f6548df8d751ced3c5bfc7c5d4f8cf4b6afdb02d2bb7e"} Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.713042 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68045f2c90cda241041f6548df8d751ced3c5bfc7c5d4f8cf4b6afdb02d2bb7e" Mar 09 09:30:03 crc kubenswrapper[4692]: I0309 09:30:03.912272 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550810-4ckmx" Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.052859 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76qw8\" (UniqueName: \"kubernetes.io/projected/4acdeb1a-b335-45c9-9ba6-22fe29675aaf-kube-api-access-76qw8\") pod \"4acdeb1a-b335-45c9-9ba6-22fe29675aaf\" (UID: \"4acdeb1a-b335-45c9-9ba6-22fe29675aaf\") " Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.056018 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4acdeb1a-b335-45c9-9ba6-22fe29675aaf-kube-api-access-76qw8" (OuterVolumeSpecName: "kube-api-access-76qw8") pod "4acdeb1a-b335-45c9-9ba6-22fe29675aaf" (UID: "4acdeb1a-b335-45c9-9ba6-22fe29675aaf"). InnerVolumeSpecName "kube-api-access-76qw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.154360 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76qw8\" (UniqueName: \"kubernetes.io/projected/4acdeb1a-b335-45c9-9ba6-22fe29675aaf-kube-api-access-76qw8\") on node \"crc\" DevicePath \"\"" Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.718133 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550810-4ckmx" event={"ID":"4acdeb1a-b335-45c9-9ba6-22fe29675aaf","Type":"ContainerDied","Data":"c34fadce9dcdc4e91a3c0c85f50d786350dbade181b8e9ace93dee79b702b598"} Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.718183 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c34fadce9dcdc4e91a3c0c85f50d786350dbade181b8e9ace93dee79b702b598" Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.718269 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550810-4ckmx" Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.964853 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550804-bcnrd"] Mar 09 09:30:04 crc kubenswrapper[4692]: I0309 09:30:04.968604 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550804-bcnrd"] Mar 09 09:30:06 crc kubenswrapper[4692]: I0309 09:30:06.081587 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1e0e894-d531-4ca0-a408-1118c2fa993b" path="/var/lib/kubelet/pods/c1e0e894-d531-4ca0-a408-1118c2fa993b/volumes" Mar 09 09:30:17 crc kubenswrapper[4692]: I0309 09:30:17.624824 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:30:17 crc kubenswrapper[4692]: I0309 09:30:17.625634 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.623665 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.624445 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.624487 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.625110 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31f9b48232438cf31701990fc67d0a37a49524e82d3bf2636f02423390a87487"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.625266 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://31f9b48232438cf31701990fc67d0a37a49524e82d3bf2636f02423390a87487" gracePeriod=600 Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.953137 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="31f9b48232438cf31701990fc67d0a37a49524e82d3bf2636f02423390a87487" exitCode=0 Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.953198 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"31f9b48232438cf31701990fc67d0a37a49524e82d3bf2636f02423390a87487"} Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.953475 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"5fa3fd1dcbe9393b3fc6c43e0491320471ea475c9f05b37486ae66af62282e34"} Mar 09 09:30:47 crc kubenswrapper[4692]: I0309 09:30:47.953495 4692 scope.go:117] "RemoveContainer" containerID="61fae0ebf34e8ab848fb0659a787a8b14e18310fadde50c12777c1d7eafc19fe" Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.808408 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-k72pq"] Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.809427 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-controller" containerID="cri-o://7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" gracePeriod=30 Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.809563 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="northd" containerID="cri-o://883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" gracePeriod=30 Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.809527 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="nbdb" containerID="cri-o://ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" gracePeriod=30 Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.809605 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" gracePeriod=30 Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.809670 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-acl-logging" containerID="cri-o://b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" gracePeriod=30 Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.809803 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-node" containerID="cri-o://833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" gracePeriod=30 Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.809839 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="sbdb" containerID="cri-o://be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" gracePeriod=30 Mar 09 09:31:50 crc kubenswrapper[4692]: I0309 09:31:50.837452 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovnkube-controller" containerID="cri-o://b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" gracePeriod=30 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.153319 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-k72pq_5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6/ovn-acl-logging/0.log" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.153924 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-k72pq_5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6/ovn-controller/0.log" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.155074 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.208805 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tk4kg"] Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209042 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="sbdb" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209060 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="sbdb" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209073 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69976e47-d9ef-4c41-8060-eb950d844750" containerName="collect-profiles" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209082 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="69976e47-d9ef-4c41-8060-eb950d844750" containerName="collect-profiles" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209095 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-acl-logging" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209103 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-acl-logging" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209114 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4acdeb1a-b335-45c9-9ba6-22fe29675aaf" containerName="oc" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209122 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4acdeb1a-b335-45c9-9ba6-22fe29675aaf" containerName="oc" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209133 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="nbdb" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209140 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="nbdb" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209154 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="northd" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209181 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="northd" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209196 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-controller" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209203 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-controller" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209215 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kubecfg-setup" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209223 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kubecfg-setup" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209234 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-ovn-metrics" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209241 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-ovn-metrics" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209253 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-node" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209260 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-node" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.209273 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovnkube-controller" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209280 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovnkube-controller" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209390 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4acdeb1a-b335-45c9-9ba6-22fe29675aaf" containerName="oc" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209402 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="nbdb" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209411 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-ovn-metrics" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209420 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="northd" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209428 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-acl-logging" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209440 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovn-controller" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209451 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="69976e47-d9ef-4c41-8060-eb950d844750" containerName="collect-profiles" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209460 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="kube-rbac-proxy-node" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209469 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="sbdb" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.209479 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerName="ovnkube-controller" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.211653 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251754 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-kubelet\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251809 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-ovn\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251844 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-env-overrides\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251860 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-log-socket\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251875 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-var-lib-openvswitch\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251910 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovn-node-metrics-cert\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251930 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-systemd\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251954 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv54g\" (UniqueName: \"kubernetes.io/projected/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-kube-api-access-pv54g\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251972 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-config\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.251993 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-openvswitch\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252016 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-systemd-units\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252034 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-slash\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252063 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-ovn-kubernetes\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252070 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252083 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-bin\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252125 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-etc-openvswitch\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252142 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-netns\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252156 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252151 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252209 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-netd\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252237 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-node-log\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252245 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-log-socket" (OuterVolumeSpecName: "log-socket") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252255 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-script-lib\") pod \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\" (UID: \"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6\") " Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252269 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-slash" (OuterVolumeSpecName: "host-slash") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252254 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252433 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252461 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252607 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252660 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252672 4692 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252689 4692 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252700 4692 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-log-socket\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252712 4692 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252720 4692 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252729 4692 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252738 4692 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-slash\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252687 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252706 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252758 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252724 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252739 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252800 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.252985 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-node-log" (OuterVolumeSpecName: "node-log") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.253104 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.257942 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-kube-api-access-pv54g" (OuterVolumeSpecName: "kube-api-access-pv54g") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "kube-api-access-pv54g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.258484 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.272734 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" (UID: "5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.274111 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2cbtq_008780c0-faad-4a8a-a6b4-4d092ff53b4a/kube-multus/0.log" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.274181 4692 generic.go:334] "Generic (PLEG): container finished" podID="008780c0-faad-4a8a-a6b4-4d092ff53b4a" containerID="c6ec87e2b80080204e209614ace6f0ad6b7cf9165ba1078af6ffd2d5214a6d6f" exitCode=2 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.274248 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2cbtq" event={"ID":"008780c0-faad-4a8a-a6b4-4d092ff53b4a","Type":"ContainerDied","Data":"c6ec87e2b80080204e209614ace6f0ad6b7cf9165ba1078af6ffd2d5214a6d6f"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.274713 4692 scope.go:117] "RemoveContainer" containerID="c6ec87e2b80080204e209614ace6f0ad6b7cf9165ba1078af6ffd2d5214a6d6f" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.279121 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-k72pq_5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6/ovn-acl-logging/0.log" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.279749 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-k72pq_5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6/ovn-controller/0.log" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280386 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" exitCode=0 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280408 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" exitCode=0 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280419 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" exitCode=0 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280426 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" exitCode=0 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280432 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" exitCode=0 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280441 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" exitCode=0 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280449 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" exitCode=143 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280456 4692 generic.go:334] "Generic (PLEG): container finished" podID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" containerID="7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" exitCode=143 Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280467 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280464 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280511 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280522 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280532 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280543 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280551 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280561 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280570 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280576 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280582 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280590 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280598 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280603 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280608 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280613 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280618 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280623 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280628 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280633 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280639 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280646 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280666 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280671 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280676 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280681 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280686 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280690 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280695 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280700 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280707 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-k72pq" event={"ID":"5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6","Type":"ContainerDied","Data":"2f2e2f093b91bba6196566d45d5dd232e73fcd4194220a9c0396b2bd8ce9431d"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280714 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280720 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280725 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280730 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280735 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280741 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280746 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280752 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280757 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.280773 4692 scope.go:117] "RemoveContainer" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.301313 4692 scope.go:117] "RemoveContainer" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.331677 4692 scope.go:117] "RemoveContainer" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.343150 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-k72pq"] Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.346130 4692 scope.go:117] "RemoveContainer" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.349679 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-k72pq"] Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353545 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-node-log\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353592 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353618 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353649 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-ovnkube-script-lib\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353670 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-env-overrides\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353690 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-etc-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353712 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-cni-netd\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353740 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-run-netns\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353763 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-var-lib-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.353999 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-run-ovn-kubernetes\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354111 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-ovnkube-config\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354232 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-systemd-units\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354317 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-ovn\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354534 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-kubelet\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354602 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-slash\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354645 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-cni-bin\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354695 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-systemd\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354752 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-log-socket\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354784 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99lql\" (UniqueName: \"kubernetes.io/projected/1f678350-d456-4868-93b3-33e04f0a0ecb-kube-api-access-99lql\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354826 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f678350-d456-4868-93b3-33e04f0a0ecb-ovn-node-metrics-cert\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354875 4692 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354893 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv54g\" (UniqueName: \"kubernetes.io/projected/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-kube-api-access-pv54g\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354907 4692 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354920 4692 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354932 4692 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.354943 4692 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.355143 4692 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.355181 4692 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.355198 4692 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.355209 4692 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-node-log\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.355333 4692 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.355352 4692 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.355365 4692 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.370106 4692 scope.go:117] "RemoveContainer" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.384898 4692 scope.go:117] "RemoveContainer" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.399534 4692 scope.go:117] "RemoveContainer" containerID="b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.416099 4692 scope.go:117] "RemoveContainer" containerID="7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.430155 4692 scope.go:117] "RemoveContainer" containerID="00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.447044 4692 scope.go:117] "RemoveContainer" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.447483 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": container with ID starting with b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b not found: ID does not exist" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.447509 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} err="failed to get container status \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": rpc error: code = NotFound desc = could not find container \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": container with ID starting with b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.447528 4692 scope.go:117] "RemoveContainer" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.447844 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": container with ID starting with be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9 not found: ID does not exist" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.447885 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} err="failed to get container status \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": rpc error: code = NotFound desc = could not find container \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": container with ID starting with be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.447898 4692 scope.go:117] "RemoveContainer" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.448274 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": container with ID starting with ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55 not found: ID does not exist" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.448319 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} err="failed to get container status \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": rpc error: code = NotFound desc = could not find container \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": container with ID starting with ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.448335 4692 scope.go:117] "RemoveContainer" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.448697 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": container with ID starting with 883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00 not found: ID does not exist" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.448726 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} err="failed to get container status \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": rpc error: code = NotFound desc = could not find container \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": container with ID starting with 883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.448746 4692 scope.go:117] "RemoveContainer" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.449179 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": container with ID starting with 1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5 not found: ID does not exist" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.449218 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} err="failed to get container status \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": rpc error: code = NotFound desc = could not find container \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": container with ID starting with 1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.449242 4692 scope.go:117] "RemoveContainer" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.449646 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": container with ID starting with 833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff not found: ID does not exist" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.449670 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} err="failed to get container status \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": rpc error: code = NotFound desc = could not find container \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": container with ID starting with 833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.449701 4692 scope.go:117] "RemoveContainer" containerID="b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.450042 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": container with ID starting with b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5 not found: ID does not exist" containerID="b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.450086 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} err="failed to get container status \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": rpc error: code = NotFound desc = could not find container \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": container with ID starting with b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.450112 4692 scope.go:117] "RemoveContainer" containerID="7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.450404 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": container with ID starting with 7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5 not found: ID does not exist" containerID="7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.450426 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} err="failed to get container status \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": rpc error: code = NotFound desc = could not find container \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": container with ID starting with 7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.450441 4692 scope.go:117] "RemoveContainer" containerID="00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955" Mar 09 09:31:51 crc kubenswrapper[4692]: E0309 09:31:51.450699 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": container with ID starting with 00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955 not found: ID does not exist" containerID="00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.450722 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} err="failed to get container status \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": rpc error: code = NotFound desc = could not find container \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": container with ID starting with 00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.450738 4692 scope.go:117] "RemoveContainer" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451000 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} err="failed to get container status \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": rpc error: code = NotFound desc = could not find container \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": container with ID starting with b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451025 4692 scope.go:117] "RemoveContainer" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451261 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} err="failed to get container status \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": rpc error: code = NotFound desc = could not find container \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": container with ID starting with be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451281 4692 scope.go:117] "RemoveContainer" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451535 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} err="failed to get container status \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": rpc error: code = NotFound desc = could not find container \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": container with ID starting with ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451552 4692 scope.go:117] "RemoveContainer" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451733 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} err="failed to get container status \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": rpc error: code = NotFound desc = could not find container \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": container with ID starting with 883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451747 4692 scope.go:117] "RemoveContainer" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451908 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} err="failed to get container status \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": rpc error: code = NotFound desc = could not find container \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": container with ID starting with 1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.451923 4692 scope.go:117] "RemoveContainer" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452127 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} err="failed to get container status \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": rpc error: code = NotFound desc = could not find container \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": container with ID starting with 833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452150 4692 scope.go:117] "RemoveContainer" containerID="b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452385 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} err="failed to get container status \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": rpc error: code = NotFound desc = could not find container \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": container with ID starting with b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452400 4692 scope.go:117] "RemoveContainer" containerID="7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452676 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} err="failed to get container status \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": rpc error: code = NotFound desc = could not find container \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": container with ID starting with 7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452692 4692 scope.go:117] "RemoveContainer" containerID="00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452926 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} err="failed to get container status \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": rpc error: code = NotFound desc = could not find container \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": container with ID starting with 00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.452951 4692 scope.go:117] "RemoveContainer" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.453228 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} err="failed to get container status \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": rpc error: code = NotFound desc = could not find container \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": container with ID starting with b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.453248 4692 scope.go:117] "RemoveContainer" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.453491 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} err="failed to get container status \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": rpc error: code = NotFound desc = could not find container \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": container with ID starting with be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.453509 4692 scope.go:117] "RemoveContainer" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.453753 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} err="failed to get container status \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": rpc error: code = NotFound desc = could not find container \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": container with ID starting with ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.453775 4692 scope.go:117] "RemoveContainer" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.454039 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} err="failed to get container status \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": rpc error: code = NotFound desc = could not find container \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": container with ID starting with 883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.454065 4692 scope.go:117] "RemoveContainer" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.454274 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} err="failed to get container status \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": rpc error: code = NotFound desc = could not find container \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": container with ID starting with 1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.454291 4692 scope.go:117] "RemoveContainer" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.454783 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} err="failed to get container status \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": rpc error: code = NotFound desc = could not find container \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": container with ID starting with 833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.454802 4692 scope.go:117] "RemoveContainer" containerID="b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455186 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} err="failed to get container status \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": rpc error: code = NotFound desc = could not find container \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": container with ID starting with b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455237 4692 scope.go:117] "RemoveContainer" containerID="7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455598 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} err="failed to get container status \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": rpc error: code = NotFound desc = could not find container \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": container with ID starting with 7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455622 4692 scope.go:117] "RemoveContainer" containerID="00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455861 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f678350-d456-4868-93b3-33e04f0a0ecb-ovn-node-metrics-cert\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455882 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} err="failed to get container status \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": rpc error: code = NotFound desc = could not find container \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": container with ID starting with 00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455893 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-node-log\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455914 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455931 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455951 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-ovnkube-script-lib\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455968 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-env-overrides\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455983 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-etc-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455997 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-cni-netd\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456017 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-run-netns\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456034 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-var-lib-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456073 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-run-ovn-kubernetes\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456088 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-ovnkube-config\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456103 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-systemd-units\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456117 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-ovn\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456142 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-kubelet\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456183 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-slash\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456198 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-cni-bin\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456215 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-systemd\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456232 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-log-socket\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456259 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99lql\" (UniqueName: \"kubernetes.io/projected/1f678350-d456-4868-93b3-33e04f0a0ecb-kube-api-access-99lql\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456534 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-node-log\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.455896 4692 scope.go:117] "RemoveContainer" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456621 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456641 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.456847 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-systemd-units\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457215 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-run-ovn-kubernetes\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457223 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-var-lib-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457239 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-ovnkube-script-lib\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457307 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-etc-openvswitch\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457317 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-cni-bin\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457341 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-cni-netd\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457371 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-env-overrides\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457399 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-kubelet\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457488 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-systemd\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457545 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-slash\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457574 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-run-ovn\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457620 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-host-run-netns\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457820 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f678350-d456-4868-93b3-33e04f0a0ecb-log-socket\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.457910 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f678350-d456-4868-93b3-33e04f0a0ecb-ovnkube-config\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.459103 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} err="failed to get container status \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": rpc error: code = NotFound desc = could not find container \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": container with ID starting with b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.459136 4692 scope.go:117] "RemoveContainer" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.459632 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f678350-d456-4868-93b3-33e04f0a0ecb-ovn-node-metrics-cert\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.460619 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} err="failed to get container status \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": rpc error: code = NotFound desc = could not find container \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": container with ID starting with be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.460659 4692 scope.go:117] "RemoveContainer" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461042 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} err="failed to get container status \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": rpc error: code = NotFound desc = could not find container \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": container with ID starting with ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461081 4692 scope.go:117] "RemoveContainer" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461404 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} err="failed to get container status \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": rpc error: code = NotFound desc = could not find container \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": container with ID starting with 883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461433 4692 scope.go:117] "RemoveContainer" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461706 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} err="failed to get container status \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": rpc error: code = NotFound desc = could not find container \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": container with ID starting with 1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461730 4692 scope.go:117] "RemoveContainer" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461950 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} err="failed to get container status \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": rpc error: code = NotFound desc = could not find container \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": container with ID starting with 833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.461968 4692 scope.go:117] "RemoveContainer" containerID="b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.462345 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5"} err="failed to get container status \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": rpc error: code = NotFound desc = could not find container \"b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5\": container with ID starting with b2a29974f0d3771a0f812db2c8f66750659a24e8ade3cde1b242b5cebed9e8f5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.462366 4692 scope.go:117] "RemoveContainer" containerID="7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.462547 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5"} err="failed to get container status \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": rpc error: code = NotFound desc = could not find container \"7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5\": container with ID starting with 7058a8f5199411f59f7e4a7c71d90b9af46846bc90d59e0096ba604d95e654b5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.462569 4692 scope.go:117] "RemoveContainer" containerID="00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.462773 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955"} err="failed to get container status \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": rpc error: code = NotFound desc = could not find container \"00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955\": container with ID starting with 00ca169a5584950784c458a9d7727993143c5276d0a2bfd9050cbe54f6c6d955 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.462813 4692 scope.go:117] "RemoveContainer" containerID="b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463028 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b"} err="failed to get container status \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": rpc error: code = NotFound desc = could not find container \"b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b\": container with ID starting with b9e5bb43fa03055fa032860b2be284dfef8e5716dc41a66b95565cf88662338b not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463049 4692 scope.go:117] "RemoveContainer" containerID="be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463263 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9"} err="failed to get container status \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": rpc error: code = NotFound desc = could not find container \"be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9\": container with ID starting with be09506fe78e566fe7ea62d6dbb7416898d1c23d67210870b9db193b26b07fb9 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463277 4692 scope.go:117] "RemoveContainer" containerID="ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463457 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55"} err="failed to get container status \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": rpc error: code = NotFound desc = could not find container \"ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55\": container with ID starting with ddfd1fe22ec940becc90e884e3ed2afd367bdd887322fc692716d8fbd9d70e55 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463475 4692 scope.go:117] "RemoveContainer" containerID="883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463783 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00"} err="failed to get container status \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": rpc error: code = NotFound desc = could not find container \"883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00\": container with ID starting with 883f167eb6a0f8839327a2d8908f6db37aca77a743092425571800d8cf183a00 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463800 4692 scope.go:117] "RemoveContainer" containerID="1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463982 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5"} err="failed to get container status \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": rpc error: code = NotFound desc = could not find container \"1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5\": container with ID starting with 1354da5cfb43fff43a2ec54c8d42fc8779684fc66ad56c03d218c60807131ea5 not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.463998 4692 scope.go:117] "RemoveContainer" containerID="833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.464358 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff"} err="failed to get container status \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": rpc error: code = NotFound desc = could not find container \"833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff\": container with ID starting with 833d1316f286215c967a319b51c12dac4b6bb2cf9bd6905ddf357ae16b3b73ff not found: ID does not exist" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.473929 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99lql\" (UniqueName: \"kubernetes.io/projected/1f678350-d456-4868-93b3-33e04f0a0ecb-kube-api-access-99lql\") pod \"ovnkube-node-tk4kg\" (UID: \"1f678350-d456-4868-93b3-33e04f0a0ecb\") " pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: I0309 09:31:51.525679 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:51 crc kubenswrapper[4692]: W0309 09:31:51.542138 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f678350_d456_4868_93b3_33e04f0a0ecb.slice/crio-90cc8012916e8c5e596276115ea74b0854489ce2b32b7cb8f39119d17371cdf1 WatchSource:0}: Error finding container 90cc8012916e8c5e596276115ea74b0854489ce2b32b7cb8f39119d17371cdf1: Status 404 returned error can't find the container with id 90cc8012916e8c5e596276115ea74b0854489ce2b32b7cb8f39119d17371cdf1 Mar 09 09:31:52 crc kubenswrapper[4692]: I0309 09:31:52.079418 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6" path="/var/lib/kubelet/pods/5f7c20da-5bc3-4e53-a93d-90faf1bdcdf6/volumes" Mar 09 09:31:52 crc kubenswrapper[4692]: I0309 09:31:52.285777 4692 generic.go:334] "Generic (PLEG): container finished" podID="1f678350-d456-4868-93b3-33e04f0a0ecb" containerID="b1b6099e2bc30250738be862004e766bf6b7df15e23cb6b24dbf781a7bdaf85c" exitCode=0 Mar 09 09:31:52 crc kubenswrapper[4692]: I0309 09:31:52.285838 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerDied","Data":"b1b6099e2bc30250738be862004e766bf6b7df15e23cb6b24dbf781a7bdaf85c"} Mar 09 09:31:52 crc kubenswrapper[4692]: I0309 09:31:52.285862 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"90cc8012916e8c5e596276115ea74b0854489ce2b32b7cb8f39119d17371cdf1"} Mar 09 09:31:52 crc kubenswrapper[4692]: I0309 09:31:52.288572 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2cbtq_008780c0-faad-4a8a-a6b4-4d092ff53b4a/kube-multus/0.log" Mar 09 09:31:52 crc kubenswrapper[4692]: I0309 09:31:52.288647 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2cbtq" event={"ID":"008780c0-faad-4a8a-a6b4-4d092ff53b4a","Type":"ContainerStarted","Data":"6c907ea3f562510c0847a9aacbef50e8877b20e2d367ea1f50748bf2fbda1aec"} Mar 09 09:31:53 crc kubenswrapper[4692]: I0309 09:31:53.298338 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"75f7b3bf190f29304de1b34eef433592f9114d5f62ab74f80cba816b5c8a861d"} Mar 09 09:31:53 crc kubenswrapper[4692]: I0309 09:31:53.298676 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"734c7e9fa51cf3a73c0aed5da1e38e9c68e418a9b030cca8138ce6fb3426d573"} Mar 09 09:31:53 crc kubenswrapper[4692]: I0309 09:31:53.298694 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"64c69c2cc8a5189d3641b93b564d6d18eb37207f6e9e22f589721ce0bb0536a9"} Mar 09 09:31:53 crc kubenswrapper[4692]: I0309 09:31:53.298705 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"756e927323697a388cb119f3a5fe89b9f9d2dc13abebffdb8b5a455d660ef0d5"} Mar 09 09:31:53 crc kubenswrapper[4692]: I0309 09:31:53.298716 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"3226a736ee355dc35939f9bfb4aba44b75113e283c5b8c1e6ee374ae6cbc0f34"} Mar 09 09:31:53 crc kubenswrapper[4692]: I0309 09:31:53.298727 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"7ba2db0102eefc5a49382a06a7d37fdaae8dec23d38a1344ad6bf3ec8b35c987"} Mar 09 09:31:55 crc kubenswrapper[4692]: I0309 09:31:55.313432 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"aee66dbb9a1b9f3cb33ad0453aa4b744e7adb56a50c9823c6bd018eb825601f9"} Mar 09 09:31:57 crc kubenswrapper[4692]: I0309 09:31:57.325284 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" event={"ID":"1f678350-d456-4868-93b3-33e04f0a0ecb","Type":"ContainerStarted","Data":"bd3b7a3992b1cd6e332440dae4c3919f75af490bdbe1bef73d2ccd9898c66601"} Mar 09 09:31:58 crc kubenswrapper[4692]: I0309 09:31:58.331353 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:58 crc kubenswrapper[4692]: I0309 09:31:58.331394 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:58 crc kubenswrapper[4692]: I0309 09:31:58.331404 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:58 crc kubenswrapper[4692]: I0309 09:31:58.364950 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:58 crc kubenswrapper[4692]: I0309 09:31:58.369051 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:31:58 crc kubenswrapper[4692]: I0309 09:31:58.398854 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" podStartSLOduration=7.398834652 podStartE2EDuration="7.398834652s" podCreationTimestamp="2026-03-09 09:31:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:31:58.36594741 +0000 UTC m=+719.190683011" watchObservedRunningTime="2026-03-09 09:31:58.398834652 +0000 UTC m=+719.223570243" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.130352 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550812-prhwl"] Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.131079 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550812-prhwl" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.133079 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.133391 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.133395 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.138348 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550812-prhwl"] Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.258937 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c7x7\" (UniqueName: \"kubernetes.io/projected/dcaa35a6-212d-4f51-bdc6-77d45a0d589c-kube-api-access-9c7x7\") pod \"auto-csr-approver-29550812-prhwl\" (UID: \"dcaa35a6-212d-4f51-bdc6-77d45a0d589c\") " pod="openshift-infra/auto-csr-approver-29550812-prhwl" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.360994 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c7x7\" (UniqueName: \"kubernetes.io/projected/dcaa35a6-212d-4f51-bdc6-77d45a0d589c-kube-api-access-9c7x7\") pod \"auto-csr-approver-29550812-prhwl\" (UID: \"dcaa35a6-212d-4f51-bdc6-77d45a0d589c\") " pod="openshift-infra/auto-csr-approver-29550812-prhwl" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.386883 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c7x7\" (UniqueName: \"kubernetes.io/projected/dcaa35a6-212d-4f51-bdc6-77d45a0d589c-kube-api-access-9c7x7\") pod \"auto-csr-approver-29550812-prhwl\" (UID: \"dcaa35a6-212d-4f51-bdc6-77d45a0d589c\") " pod="openshift-infra/auto-csr-approver-29550812-prhwl" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.453085 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550812-prhwl" Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.635731 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550812-prhwl"] Mar 09 09:32:00 crc kubenswrapper[4692]: W0309 09:32:00.637919 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddcaa35a6_212d_4f51_bdc6_77d45a0d589c.slice/crio-6c01de4fa5209b33a43191c4bc93362a18b675c0d8c8d7e4519716bdd8b8c85e WatchSource:0}: Error finding container 6c01de4fa5209b33a43191c4bc93362a18b675c0d8c8d7e4519716bdd8b8c85e: Status 404 returned error can't find the container with id 6c01de4fa5209b33a43191c4bc93362a18b675c0d8c8d7e4519716bdd8b8c85e Mar 09 09:32:00 crc kubenswrapper[4692]: I0309 09:32:00.914644 4692 scope.go:117] "RemoveContainer" containerID="714119e2a48d8ee67c9b24a147de76be865fafb71e6d7e25d7d174f99ab0974e" Mar 09 09:32:01 crc kubenswrapper[4692]: I0309 09:32:01.347454 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550812-prhwl" event={"ID":"dcaa35a6-212d-4f51-bdc6-77d45a0d589c","Type":"ContainerStarted","Data":"6c01de4fa5209b33a43191c4bc93362a18b675c0d8c8d7e4519716bdd8b8c85e"} Mar 09 09:32:02 crc kubenswrapper[4692]: I0309 09:32:02.352882 4692 generic.go:334] "Generic (PLEG): container finished" podID="dcaa35a6-212d-4f51-bdc6-77d45a0d589c" containerID="7322dbc30ea39903ba46c40fb451a821a0ca196163a623f7eaf62742fb37905d" exitCode=0 Mar 09 09:32:02 crc kubenswrapper[4692]: I0309 09:32:02.352978 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550812-prhwl" event={"ID":"dcaa35a6-212d-4f51-bdc6-77d45a0d589c","Type":"ContainerDied","Data":"7322dbc30ea39903ba46c40fb451a821a0ca196163a623f7eaf62742fb37905d"} Mar 09 09:32:03 crc kubenswrapper[4692]: I0309 09:32:03.594392 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550812-prhwl" Mar 09 09:32:03 crc kubenswrapper[4692]: I0309 09:32:03.714426 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c7x7\" (UniqueName: \"kubernetes.io/projected/dcaa35a6-212d-4f51-bdc6-77d45a0d589c-kube-api-access-9c7x7\") pod \"dcaa35a6-212d-4f51-bdc6-77d45a0d589c\" (UID: \"dcaa35a6-212d-4f51-bdc6-77d45a0d589c\") " Mar 09 09:32:03 crc kubenswrapper[4692]: I0309 09:32:03.720869 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcaa35a6-212d-4f51-bdc6-77d45a0d589c-kube-api-access-9c7x7" (OuterVolumeSpecName: "kube-api-access-9c7x7") pod "dcaa35a6-212d-4f51-bdc6-77d45a0d589c" (UID: "dcaa35a6-212d-4f51-bdc6-77d45a0d589c"). InnerVolumeSpecName "kube-api-access-9c7x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:32:03 crc kubenswrapper[4692]: I0309 09:32:03.816825 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c7x7\" (UniqueName: \"kubernetes.io/projected/dcaa35a6-212d-4f51-bdc6-77d45a0d589c-kube-api-access-9c7x7\") on node \"crc\" DevicePath \"\"" Mar 09 09:32:04 crc kubenswrapper[4692]: I0309 09:32:04.363821 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550812-prhwl" event={"ID":"dcaa35a6-212d-4f51-bdc6-77d45a0d589c","Type":"ContainerDied","Data":"6c01de4fa5209b33a43191c4bc93362a18b675c0d8c8d7e4519716bdd8b8c85e"} Mar 09 09:32:04 crc kubenswrapper[4692]: I0309 09:32:04.364452 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c01de4fa5209b33a43191c4bc93362a18b675c0d8c8d7e4519716bdd8b8c85e" Mar 09 09:32:04 crc kubenswrapper[4692]: I0309 09:32:04.364095 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550812-prhwl" Mar 09 09:32:04 crc kubenswrapper[4692]: I0309 09:32:04.661412 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550806-99ld4"] Mar 09 09:32:04 crc kubenswrapper[4692]: I0309 09:32:04.665355 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550806-99ld4"] Mar 09 09:32:06 crc kubenswrapper[4692]: I0309 09:32:06.079051 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f98f499-bbc9-4ae7-b470-dfe2a771127e" path="/var/lib/kubelet/pods/9f98f499-bbc9-4ae7-b470-dfe2a771127e/volumes" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.641138 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg"] Mar 09 09:32:14 crc kubenswrapper[4692]: E0309 09:32:14.642010 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcaa35a6-212d-4f51-bdc6-77d45a0d589c" containerName="oc" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.642026 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcaa35a6-212d-4f51-bdc6-77d45a0d589c" containerName="oc" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.642150 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcaa35a6-212d-4f51-bdc6-77d45a0d589c" containerName="oc" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.643019 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.643852 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.643951 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdtbq\" (UniqueName: \"kubernetes.io/projected/e6640a53-9679-4b36-a32d-c0fe67ecfe26-kube-api-access-kdtbq\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.644127 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.648260 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.649266 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg"] Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.744826 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdtbq\" (UniqueName: \"kubernetes.io/projected/e6640a53-9679-4b36-a32d-c0fe67ecfe26-kube-api-access-kdtbq\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.744905 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.744966 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.745581 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.745718 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.763972 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdtbq\" (UniqueName: \"kubernetes.io/projected/e6640a53-9679-4b36-a32d-c0fe67ecfe26-kube-api-access-kdtbq\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:14 crc kubenswrapper[4692]: I0309 09:32:14.962994 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:15 crc kubenswrapper[4692]: I0309 09:32:15.364694 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg"] Mar 09 09:32:15 crc kubenswrapper[4692]: I0309 09:32:15.422582 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" event={"ID":"e6640a53-9679-4b36-a32d-c0fe67ecfe26","Type":"ContainerStarted","Data":"f49ad3cc0776b43231aef0bd858461a1beaf91a153f83f5c0e5c92199ad3bcad"} Mar 09 09:32:16 crc kubenswrapper[4692]: I0309 09:32:16.439910 4692 generic.go:334] "Generic (PLEG): container finished" podID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerID="f0a02da98761ef6a7551210376a2f092e3ef2d7c577225896de90e46e7c7a0a5" exitCode=0 Mar 09 09:32:16 crc kubenswrapper[4692]: I0309 09:32:16.440054 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" event={"ID":"e6640a53-9679-4b36-a32d-c0fe67ecfe26","Type":"ContainerDied","Data":"f0a02da98761ef6a7551210376a2f092e3ef2d7c577225896de90e46e7c7a0a5"} Mar 09 09:32:16 crc kubenswrapper[4692]: I0309 09:32:16.985939 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bnpkl"] Mar 09 09:32:16 crc kubenswrapper[4692]: I0309 09:32:16.987076 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.006376 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnpkl"] Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.072304 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-utilities\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.072618 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xblt8\" (UniqueName: \"kubernetes.io/projected/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-kube-api-access-xblt8\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.072707 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-catalog-content\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.174333 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xblt8\" (UniqueName: \"kubernetes.io/projected/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-kube-api-access-xblt8\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.174400 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-catalog-content\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.174441 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-utilities\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.174894 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-utilities\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.175511 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-catalog-content\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.195103 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xblt8\" (UniqueName: \"kubernetes.io/projected/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-kube-api-access-xblt8\") pod \"redhat-operators-bnpkl\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.302368 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:17 crc kubenswrapper[4692]: I0309 09:32:17.732077 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnpkl"] Mar 09 09:32:18 crc kubenswrapper[4692]: I0309 09:32:18.461556 4692 generic.go:334] "Generic (PLEG): container finished" podID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerID="f3902effbf0d7f969cd4bc674ea448cac948c8e0a983ccd5b68ce5d48e0934a9" exitCode=0 Mar 09 09:32:18 crc kubenswrapper[4692]: I0309 09:32:18.461620 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnpkl" event={"ID":"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb","Type":"ContainerDied","Data":"f3902effbf0d7f969cd4bc674ea448cac948c8e0a983ccd5b68ce5d48e0934a9"} Mar 09 09:32:18 crc kubenswrapper[4692]: I0309 09:32:18.461662 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnpkl" event={"ID":"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb","Type":"ContainerStarted","Data":"ab4efab814c7b5ff3596147845d79ade25ab9d9a873636031db293bbbc291dde"} Mar 09 09:32:18 crc kubenswrapper[4692]: I0309 09:32:18.463792 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" event={"ID":"e6640a53-9679-4b36-a32d-c0fe67ecfe26","Type":"ContainerDied","Data":"b57a7f55f8da4f8128581a5d37cb2921db54022fe77bab84c7e985caf415bed9"} Mar 09 09:32:18 crc kubenswrapper[4692]: I0309 09:32:18.463701 4692 generic.go:334] "Generic (PLEG): container finished" podID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerID="b57a7f55f8da4f8128581a5d37cb2921db54022fe77bab84c7e985caf415bed9" exitCode=0 Mar 09 09:32:18 crc kubenswrapper[4692]: I0309 09:32:18.562327 4692 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 09 09:32:19 crc kubenswrapper[4692]: I0309 09:32:19.473058 4692 generic.go:334] "Generic (PLEG): container finished" podID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerID="4f79ade32185f1ce7146e7e8589da3cb762dc39085771a9f0cbfe7d8ce91d736" exitCode=0 Mar 09 09:32:19 crc kubenswrapper[4692]: I0309 09:32:19.473186 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" event={"ID":"e6640a53-9679-4b36-a32d-c0fe67ecfe26","Type":"ContainerDied","Data":"4f79ade32185f1ce7146e7e8589da3cb762dc39085771a9f0cbfe7d8ce91d736"} Mar 09 09:32:19 crc kubenswrapper[4692]: I0309 09:32:19.475221 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnpkl" event={"ID":"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb","Type":"ContainerStarted","Data":"2a8880cfdd4fd8f7ee4a79361b9064d973b94cbb2890eede229f76543c033544"} Mar 09 09:32:20 crc kubenswrapper[4692]: I0309 09:32:20.869350 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.025825 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-bundle\") pod \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.025888 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-util\") pod \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.025947 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdtbq\" (UniqueName: \"kubernetes.io/projected/e6640a53-9679-4b36-a32d-c0fe67ecfe26-kube-api-access-kdtbq\") pod \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\" (UID: \"e6640a53-9679-4b36-a32d-c0fe67ecfe26\") " Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.026886 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-bundle" (OuterVolumeSpecName: "bundle") pod "e6640a53-9679-4b36-a32d-c0fe67ecfe26" (UID: "e6640a53-9679-4b36-a32d-c0fe67ecfe26"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.032659 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6640a53-9679-4b36-a32d-c0fe67ecfe26-kube-api-access-kdtbq" (OuterVolumeSpecName: "kube-api-access-kdtbq") pod "e6640a53-9679-4b36-a32d-c0fe67ecfe26" (UID: "e6640a53-9679-4b36-a32d-c0fe67ecfe26"). InnerVolumeSpecName "kube-api-access-kdtbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.036363 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-util" (OuterVolumeSpecName: "util") pod "e6640a53-9679-4b36-a32d-c0fe67ecfe26" (UID: "e6640a53-9679-4b36-a32d-c0fe67ecfe26"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.126941 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdtbq\" (UniqueName: \"kubernetes.io/projected/e6640a53-9679-4b36-a32d-c0fe67ecfe26-kube-api-access-kdtbq\") on node \"crc\" DevicePath \"\"" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.126991 4692 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.127016 4692 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6640a53-9679-4b36-a32d-c0fe67ecfe26-util\") on node \"crc\" DevicePath \"\"" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.489856 4692 generic.go:334] "Generic (PLEG): container finished" podID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerID="2a8880cfdd4fd8f7ee4a79361b9064d973b94cbb2890eede229f76543c033544" exitCode=0 Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.489955 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnpkl" event={"ID":"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb","Type":"ContainerDied","Data":"2a8880cfdd4fd8f7ee4a79361b9064d973b94cbb2890eede229f76543c033544"} Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.497542 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" event={"ID":"e6640a53-9679-4b36-a32d-c0fe67ecfe26","Type":"ContainerDied","Data":"f49ad3cc0776b43231aef0bd858461a1beaf91a153f83f5c0e5c92199ad3bcad"} Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.497709 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f49ad3cc0776b43231aef0bd858461a1beaf91a153f83f5c0e5c92199ad3bcad" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.497863 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg" Mar 09 09:32:21 crc kubenswrapper[4692]: I0309 09:32:21.551602 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tk4kg" Mar 09 09:32:22 crc kubenswrapper[4692]: I0309 09:32:22.505856 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnpkl" event={"ID":"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb","Type":"ContainerStarted","Data":"76fbd4cfcca7f1bead626ec4f1da32a939c7a3c31327ee68350e09790cd2a755"} Mar 09 09:32:22 crc kubenswrapper[4692]: I0309 09:32:22.531584 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bnpkl" podStartSLOduration=3.086512289 podStartE2EDuration="6.531565283s" podCreationTimestamp="2026-03-09 09:32:16 +0000 UTC" firstStartedPulling="2026-03-09 09:32:18.463092219 +0000 UTC m=+739.287827800" lastFinishedPulling="2026-03-09 09:32:21.908145213 +0000 UTC m=+742.732880794" observedRunningTime="2026-03-09 09:32:22.529671838 +0000 UTC m=+743.354407429" watchObservedRunningTime="2026-03-09 09:32:22.531565283 +0000 UTC m=+743.356300884" Mar 09 09:32:27 crc kubenswrapper[4692]: I0309 09:32:27.302816 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:27 crc kubenswrapper[4692]: I0309 09:32:27.304241 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:28 crc kubenswrapper[4692]: I0309 09:32:28.348152 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bnpkl" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="registry-server" probeResult="failure" output=< Mar 09 09:32:28 crc kubenswrapper[4692]: timeout: failed to connect service ":50051" within 1s Mar 09 09:32:28 crc kubenswrapper[4692]: > Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.791050 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l"] Mar 09 09:32:31 crc kubenswrapper[4692]: E0309 09:32:31.791911 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerName="pull" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.791927 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerName="pull" Mar 09 09:32:31 crc kubenswrapper[4692]: E0309 09:32:31.791959 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerName="util" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.791967 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerName="util" Mar 09 09:32:31 crc kubenswrapper[4692]: E0309 09:32:31.791984 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerName="extract" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.791993 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerName="extract" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.792263 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6640a53-9679-4b36-a32d-c0fe67ecfe26" containerName="extract" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.792914 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.801138 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.801265 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.801175 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-r9dxg" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.801516 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.801737 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.872080 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-apiservice-cert\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.872197 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-webhook-cert\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.872355 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxk5v\" (UniqueName: \"kubernetes.io/projected/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-kube-api-access-zxk5v\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.873449 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l"] Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.974111 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxk5v\" (UniqueName: \"kubernetes.io/projected/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-kube-api-access-zxk5v\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.974209 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-apiservice-cert\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.974259 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-webhook-cert\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.994709 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-apiservice-cert\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:31 crc kubenswrapper[4692]: I0309 09:32:31.994811 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-webhook-cert\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.002117 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxk5v\" (UniqueName: \"kubernetes.io/projected/aeb7fbbf-aa00-4666-aa7f-55abaac713c4-kube-api-access-zxk5v\") pod \"metallb-operator-controller-manager-566dcbb64d-j485l\" (UID: \"aeb7fbbf-aa00-4666-aa7f-55abaac713c4\") " pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.121536 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.174471 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d76797467-xnptd"] Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.175098 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.184590 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.184679 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-lv4gl" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.184813 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.252793 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d76797467-xnptd"] Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.277764 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn4ms\" (UniqueName: \"kubernetes.io/projected/a9359a7b-47b3-4747-aa40-2633db5e907a-kube-api-access-dn4ms\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.277823 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9359a7b-47b3-4747-aa40-2633db5e907a-webhook-cert\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.277891 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9359a7b-47b3-4747-aa40-2633db5e907a-apiservice-cert\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.383568 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn4ms\" (UniqueName: \"kubernetes.io/projected/a9359a7b-47b3-4747-aa40-2633db5e907a-kube-api-access-dn4ms\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.383636 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9359a7b-47b3-4747-aa40-2633db5e907a-webhook-cert\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.383695 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9359a7b-47b3-4747-aa40-2633db5e907a-apiservice-cert\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.390190 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9359a7b-47b3-4747-aa40-2633db5e907a-webhook-cert\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.390267 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9359a7b-47b3-4747-aa40-2633db5e907a-apiservice-cert\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.453051 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn4ms\" (UniqueName: \"kubernetes.io/projected/a9359a7b-47b3-4747-aa40-2633db5e907a-kube-api-access-dn4ms\") pod \"metallb-operator-webhook-server-5d76797467-xnptd\" (UID: \"a9359a7b-47b3-4747-aa40-2633db5e907a\") " pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.505438 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:32 crc kubenswrapper[4692]: I0309 09:32:32.880113 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l"] Mar 09 09:32:33 crc kubenswrapper[4692]: I0309 09:32:33.055039 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5d76797467-xnptd"] Mar 09 09:32:33 crc kubenswrapper[4692]: I0309 09:32:33.561952 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" event={"ID":"a9359a7b-47b3-4747-aa40-2633db5e907a","Type":"ContainerStarted","Data":"edff1c2a6d55ceb76d8a633eb923164459b247c25ec4e75cf77f03fee5c683ce"} Mar 09 09:32:33 crc kubenswrapper[4692]: I0309 09:32:33.562782 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" event={"ID":"aeb7fbbf-aa00-4666-aa7f-55abaac713c4","Type":"ContainerStarted","Data":"bf7e114c3c37526ba89b1bd7ee9a5e63f6c5d5c65ee77c542ad229ddce60fe85"} Mar 09 09:32:37 crc kubenswrapper[4692]: I0309 09:32:37.370507 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:37 crc kubenswrapper[4692]: I0309 09:32:37.466895 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:39 crc kubenswrapper[4692]: I0309 09:32:39.185441 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnpkl"] Mar 09 09:32:39 crc kubenswrapper[4692]: I0309 09:32:39.186507 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bnpkl" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="registry-server" containerID="cri-o://76fbd4cfcca7f1bead626ec4f1da32a939c7a3c31327ee68350e09790cd2a755" gracePeriod=2 Mar 09 09:32:39 crc kubenswrapper[4692]: I0309 09:32:39.873488 4692 generic.go:334] "Generic (PLEG): container finished" podID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerID="76fbd4cfcca7f1bead626ec4f1da32a939c7a3c31327ee68350e09790cd2a755" exitCode=0 Mar 09 09:32:39 crc kubenswrapper[4692]: I0309 09:32:39.873564 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnpkl" event={"ID":"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb","Type":"ContainerDied","Data":"76fbd4cfcca7f1bead626ec4f1da32a939c7a3c31327ee68350e09790cd2a755"} Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.469172 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.538394 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-utilities\") pod \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.538515 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xblt8\" (UniqueName: \"kubernetes.io/projected/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-kube-api-access-xblt8\") pod \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.538545 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-catalog-content\") pod \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\" (UID: \"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb\") " Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.539705 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-utilities" (OuterVolumeSpecName: "utilities") pod "4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" (UID: "4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.543431 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-kube-api-access-xblt8" (OuterVolumeSpecName: "kube-api-access-xblt8") pod "4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" (UID: "4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb"). InnerVolumeSpecName "kube-api-access-xblt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.639769 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xblt8\" (UniqueName: \"kubernetes.io/projected/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-kube-api-access-xblt8\") on node \"crc\" DevicePath \"\"" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.639826 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.666417 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" (UID: "4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.740683 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.904650 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" event={"ID":"aeb7fbbf-aa00-4666-aa7f-55abaac713c4","Type":"ContainerStarted","Data":"778b7aa30b83fb4215b92b80a8e44df994e7e0bacd9aedbbe0c97f29d0ff80d4"} Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.904964 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.906303 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" event={"ID":"a9359a7b-47b3-4747-aa40-2633db5e907a","Type":"ContainerStarted","Data":"0b0cf79e33337d3ca3700bc865b55b4d825b84cc27871931a58e691b272d8c32"} Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.906430 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.908508 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnpkl" event={"ID":"4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb","Type":"ContainerDied","Data":"ab4efab814c7b5ff3596147845d79ade25ab9d9a873636031db293bbbc291dde"} Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.908546 4692 scope.go:117] "RemoveContainer" containerID="76fbd4cfcca7f1bead626ec4f1da32a939c7a3c31327ee68350e09790cd2a755" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.908577 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnpkl" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.927675 4692 scope.go:117] "RemoveContainer" containerID="2a8880cfdd4fd8f7ee4a79361b9064d973b94cbb2890eede229f76543c033544" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.928340 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" podStartSLOduration=2.3523254749999998 podStartE2EDuration="14.928327501s" podCreationTimestamp="2026-03-09 09:32:31 +0000 UTC" firstStartedPulling="2026-03-09 09:32:32.895416097 +0000 UTC m=+753.720151668" lastFinishedPulling="2026-03-09 09:32:45.471418113 +0000 UTC m=+766.296153694" observedRunningTime="2026-03-09 09:32:45.92690931 +0000 UTC m=+766.751644911" watchObservedRunningTime="2026-03-09 09:32:45.928327501 +0000 UTC m=+766.753063082" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.945519 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnpkl"] Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.948768 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bnpkl"] Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.960389 4692 scope.go:117] "RemoveContainer" containerID="f3902effbf0d7f969cd4bc674ea448cac948c8e0a983ccd5b68ce5d48e0934a9" Mar 09 09:32:45 crc kubenswrapper[4692]: I0309 09:32:45.964611 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" podStartSLOduration=1.531876698 podStartE2EDuration="13.964587082s" podCreationTimestamp="2026-03-09 09:32:32 +0000 UTC" firstStartedPulling="2026-03-09 09:32:33.069518081 +0000 UTC m=+753.894253662" lastFinishedPulling="2026-03-09 09:32:45.502228465 +0000 UTC m=+766.326964046" observedRunningTime="2026-03-09 09:32:45.962354497 +0000 UTC m=+766.787090098" watchObservedRunningTime="2026-03-09 09:32:45.964587082 +0000 UTC m=+766.789322673" Mar 09 09:32:46 crc kubenswrapper[4692]: I0309 09:32:46.085756 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" path="/var/lib/kubelet/pods/4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb/volumes" Mar 09 09:32:47 crc kubenswrapper[4692]: I0309 09:32:47.624099 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:32:47 crc kubenswrapper[4692]: I0309 09:32:47.624484 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:33:00 crc kubenswrapper[4692]: I0309 09:33:00.973635 4692 scope.go:117] "RemoveContainer" containerID="162b6ea9823b8b3b412a77b68ee22dcc0eb9b5dcf4a5abf93fd06d9d6832a578" Mar 09 09:33:02 crc kubenswrapper[4692]: I0309 09:33:02.512430 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5d76797467-xnptd" Mar 09 09:33:17 crc kubenswrapper[4692]: I0309 09:33:17.623733 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:33:17 crc kubenswrapper[4692]: I0309 09:33:17.624054 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.124627 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-566dcbb64d-j485l" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.886873 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-5b8n8"] Mar 09 09:33:22 crc kubenswrapper[4692]: E0309 09:33:22.887192 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="registry-server" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.887209 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="registry-server" Mar 09 09:33:22 crc kubenswrapper[4692]: E0309 09:33:22.887236 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="extract-utilities" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.887244 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="extract-utilities" Mar 09 09:33:22 crc kubenswrapper[4692]: E0309 09:33:22.887255 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="extract-content" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.887262 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="extract-content" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.887369 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a02ffdd-5bc7-4d07-bae8-b10b3e3be9cb" containerName="registry-server" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.889613 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.893937 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq"] Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.894853 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.894961 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.895076 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.895324 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-fq647" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.896724 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.915886 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq"] Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.991947 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6nhdr"] Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.992856 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6nhdr" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.994666 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.995087 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dj4zr" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.995128 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 09 09:33:22 crc kubenswrapper[4692]: I0309 09:33:22.995333 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007109 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-reloader\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007182 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-frr-conf\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007219 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvdxx\" (UniqueName: \"kubernetes.io/projected/32f4683a-ccb2-44de-a220-6e279f138d99-kube-api-access-qvdxx\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007243 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/32f4683a-ccb2-44de-a220-6e279f138d99-frr-startup\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007269 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-frr-sockets\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007301 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1e68ce3-7af3-4509-a31e-81f42f223aa9-cert\") pod \"frr-k8s-webhook-server-7f989f654f-n6cmq\" (UID: \"c1e68ce3-7af3-4509-a31e-81f42f223aa9\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007358 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-metrics\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007387 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32f4683a-ccb2-44de-a220-6e279f138d99-metrics-certs\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.007420 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz59q\" (UniqueName: \"kubernetes.io/projected/c1e68ce3-7af3-4509-a31e-81f42f223aa9-kube-api-access-fz59q\") pod \"frr-k8s-webhook-server-7f989f654f-n6cmq\" (UID: \"c1e68ce3-7af3-4509-a31e-81f42f223aa9\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.008710 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-86ddb6bd46-fhbtn"] Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.009875 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.013703 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.031778 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-fhbtn"] Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109310 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz59q\" (UniqueName: \"kubernetes.io/projected/c1e68ce3-7af3-4509-a31e-81f42f223aa9-kube-api-access-fz59q\") pod \"frr-k8s-webhook-server-7f989f654f-n6cmq\" (UID: \"c1e68ce3-7af3-4509-a31e-81f42f223aa9\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109350 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-reloader\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109374 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2j98\" (UniqueName: \"kubernetes.io/projected/48590eda-9d17-4bed-9911-4aad14e46289-kube-api-access-x2j98\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109399 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-frr-conf\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109438 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/32f4683a-ccb2-44de-a220-6e279f138d99-frr-startup\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109456 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvdxx\" (UniqueName: \"kubernetes.io/projected/32f4683a-ccb2-44de-a220-6e279f138d99-kube-api-access-qvdxx\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109474 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-frr-sockets\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109498 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109533 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48590eda-9d17-4bed-9911-4aad14e46289-cert\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109553 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1e68ce3-7af3-4509-a31e-81f42f223aa9-cert\") pod \"frr-k8s-webhook-server-7f989f654f-n6cmq\" (UID: \"c1e68ce3-7af3-4509-a31e-81f42f223aa9\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109607 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-metrics-certs\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109647 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8b748f8c-3824-4228-bec4-634e2ed9f324-metallb-excludel2\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109670 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-metrics\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109696 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32f4683a-ccb2-44de-a220-6e279f138d99-metrics-certs\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109769 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48590eda-9d17-4bed-9911-4aad14e46289-metrics-certs\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.109817 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww544\" (UniqueName: \"kubernetes.io/projected/8b748f8c-3824-4228-bec4-634e2ed9f324-kube-api-access-ww544\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.111280 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-reloader\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.111519 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-frr-conf\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.112876 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/32f4683a-ccb2-44de-a220-6e279f138d99-frr-startup\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.113453 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-frr-sockets\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.113928 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/32f4683a-ccb2-44de-a220-6e279f138d99-metrics\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: E0309 09:33:23.114073 4692 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Mar 09 09:33:23 crc kubenswrapper[4692]: E0309 09:33:23.114116 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/32f4683a-ccb2-44de-a220-6e279f138d99-metrics-certs podName:32f4683a-ccb2-44de-a220-6e279f138d99 nodeName:}" failed. No retries permitted until 2026-03-09 09:33:23.614101428 +0000 UTC m=+804.438836999 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/32f4683a-ccb2-44de-a220-6e279f138d99-metrics-certs") pod "frr-k8s-5b8n8" (UID: "32f4683a-ccb2-44de-a220-6e279f138d99") : secret "frr-k8s-certs-secret" not found Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.115594 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c1e68ce3-7af3-4509-a31e-81f42f223aa9-cert\") pod \"frr-k8s-webhook-server-7f989f654f-n6cmq\" (UID: \"c1e68ce3-7af3-4509-a31e-81f42f223aa9\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.131069 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz59q\" (UniqueName: \"kubernetes.io/projected/c1e68ce3-7af3-4509-a31e-81f42f223aa9-kube-api-access-fz59q\") pod \"frr-k8s-webhook-server-7f989f654f-n6cmq\" (UID: \"c1e68ce3-7af3-4509-a31e-81f42f223aa9\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.139441 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvdxx\" (UniqueName: \"kubernetes.io/projected/32f4683a-ccb2-44de-a220-6e279f138d99-kube-api-access-qvdxx\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.210780 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2j98\" (UniqueName: \"kubernetes.io/projected/48590eda-9d17-4bed-9911-4aad14e46289-kube-api-access-x2j98\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.210851 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.210875 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48590eda-9d17-4bed-9911-4aad14e46289-cert\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.210916 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-metrics-certs\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.210942 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8b748f8c-3824-4228-bec4-634e2ed9f324-metallb-excludel2\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: E0309 09:33:23.210980 4692 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 09 09:33:23 crc kubenswrapper[4692]: E0309 09:33:23.211096 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist podName:8b748f8c-3824-4228-bec4-634e2ed9f324 nodeName:}" failed. No retries permitted until 2026-03-09 09:33:23.711063622 +0000 UTC m=+804.535799293 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist") pod "speaker-6nhdr" (UID: "8b748f8c-3824-4228-bec4-634e2ed9f324") : secret "metallb-memberlist" not found Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.210986 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48590eda-9d17-4bed-9911-4aad14e46289-metrics-certs\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.211177 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww544\" (UniqueName: \"kubernetes.io/projected/8b748f8c-3824-4228-bec4-634e2ed9f324-kube-api-access-ww544\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.211737 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8b748f8c-3824-4228-bec4-634e2ed9f324-metallb-excludel2\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.212912 4692 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.214944 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-metrics-certs\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.215086 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/48590eda-9d17-4bed-9911-4aad14e46289-metrics-certs\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.221911 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.224665 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48590eda-9d17-4bed-9911-4aad14e46289-cert\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.229332 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww544\" (UniqueName: \"kubernetes.io/projected/8b748f8c-3824-4228-bec4-634e2ed9f324-kube-api-access-ww544\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.240067 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2j98\" (UniqueName: \"kubernetes.io/projected/48590eda-9d17-4bed-9911-4aad14e46289-kube-api-access-x2j98\") pod \"controller-86ddb6bd46-fhbtn\" (UID: \"48590eda-9d17-4bed-9911-4aad14e46289\") " pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.324542 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.451414 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq"] Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.470180 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.547464 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-fhbtn"] Mar 09 09:33:23 crc kubenswrapper[4692]: W0309 09:33:23.553398 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48590eda_9d17_4bed_9911_4aad14e46289.slice/crio-5732638cf1173f46c4c0d8cc49eb7f5960cd05ff41fbd2721901da39bfbc38d9 WatchSource:0}: Error finding container 5732638cf1173f46c4c0d8cc49eb7f5960cd05ff41fbd2721901da39bfbc38d9: Status 404 returned error can't find the container with id 5732638cf1173f46c4c0d8cc49eb7f5960cd05ff41fbd2721901da39bfbc38d9 Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.617804 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32f4683a-ccb2-44de-a220-6e279f138d99-metrics-certs\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.624063 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/32f4683a-ccb2-44de-a220-6e279f138d99-metrics-certs\") pod \"frr-k8s-5b8n8\" (UID: \"32f4683a-ccb2-44de-a220-6e279f138d99\") " pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.719054 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:23 crc kubenswrapper[4692]: E0309 09:33:23.719356 4692 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 09 09:33:23 crc kubenswrapper[4692]: E0309 09:33:23.719454 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist podName:8b748f8c-3824-4228-bec4-634e2ed9f324 nodeName:}" failed. No retries permitted until 2026-03-09 09:33:24.719433313 +0000 UTC m=+805.544168894 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist") pod "speaker-6nhdr" (UID: "8b748f8c-3824-4228-bec4-634e2ed9f324") : secret "metallb-memberlist" not found Mar 09 09:33:23 crc kubenswrapper[4692]: I0309 09:33:23.815303 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:24 crc kubenswrapper[4692]: I0309 09:33:24.115182 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" event={"ID":"c1e68ce3-7af3-4509-a31e-81f42f223aa9","Type":"ContainerStarted","Data":"5560cfe448473be2192bd9dfb8e5b9e0b206f086b12c95af0df77d0f45e12d77"} Mar 09 09:33:24 crc kubenswrapper[4692]: I0309 09:33:24.116246 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerStarted","Data":"1ca0414eae6e930bb1ae351531de9b918fc3f805dda16319cf99eba19a53cfa2"} Mar 09 09:33:24 crc kubenswrapper[4692]: I0309 09:33:24.117498 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-fhbtn" event={"ID":"48590eda-9d17-4bed-9911-4aad14e46289","Type":"ContainerStarted","Data":"b1d2a5e91ba64ee02781073e8726aed686eb336b3d26d6be58117645fee70586"} Mar 09 09:33:24 crc kubenswrapper[4692]: I0309 09:33:24.117524 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-fhbtn" event={"ID":"48590eda-9d17-4bed-9911-4aad14e46289","Type":"ContainerStarted","Data":"5732638cf1173f46c4c0d8cc49eb7f5960cd05ff41fbd2721901da39bfbc38d9"} Mar 09 09:33:24 crc kubenswrapper[4692]: I0309 09:33:24.731565 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:24 crc kubenswrapper[4692]: I0309 09:33:24.736759 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8b748f8c-3824-4228-bec4-634e2ed9f324-memberlist\") pod \"speaker-6nhdr\" (UID: \"8b748f8c-3824-4228-bec4-634e2ed9f324\") " pod="metallb-system/speaker-6nhdr" Mar 09 09:33:24 crc kubenswrapper[4692]: I0309 09:33:24.806414 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6nhdr" Mar 09 09:33:24 crc kubenswrapper[4692]: W0309 09:33:24.829603 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b748f8c_3824_4228_bec4_634e2ed9f324.slice/crio-10d040425d5843cc67468b76ec37e42c0b7e4232a248e7f650925f8fc5fc7c5c WatchSource:0}: Error finding container 10d040425d5843cc67468b76ec37e42c0b7e4232a248e7f650925f8fc5fc7c5c: Status 404 returned error can't find the container with id 10d040425d5843cc67468b76ec37e42c0b7e4232a248e7f650925f8fc5fc7c5c Mar 09 09:33:25 crc kubenswrapper[4692]: I0309 09:33:25.124538 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6nhdr" event={"ID":"8b748f8c-3824-4228-bec4-634e2ed9f324","Type":"ContainerStarted","Data":"a7f4a94933a4c35bcb2977bbf7a040a54529bd8728c1fb3dc675ad1ca2d447c7"} Mar 09 09:33:25 crc kubenswrapper[4692]: I0309 09:33:25.124677 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6nhdr" event={"ID":"8b748f8c-3824-4228-bec4-634e2ed9f324","Type":"ContainerStarted","Data":"10d040425d5843cc67468b76ec37e42c0b7e4232a248e7f650925f8fc5fc7c5c"} Mar 09 09:33:29 crc kubenswrapper[4692]: I0309 09:33:29.152149 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6nhdr" event={"ID":"8b748f8c-3824-4228-bec4-634e2ed9f324","Type":"ContainerStarted","Data":"51fe144ee4a02d86809c8feffe95978659774a76d09e55f00caf04aba2a7a864"} Mar 09 09:33:29 crc kubenswrapper[4692]: I0309 09:33:29.152823 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6nhdr" Mar 09 09:33:29 crc kubenswrapper[4692]: I0309 09:33:29.154531 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-fhbtn" event={"ID":"48590eda-9d17-4bed-9911-4aad14e46289","Type":"ContainerStarted","Data":"daaad29aa06d834889fb4eae8ced397e3c98cbd497493956b3b8ac6a88a9f6ec"} Mar 09 09:33:29 crc kubenswrapper[4692]: I0309 09:33:29.154646 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:29 crc kubenswrapper[4692]: I0309 09:33:29.217807 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-86ddb6bd46-fhbtn" podStartSLOduration=2.926351526 podStartE2EDuration="7.217787749s" podCreationTimestamp="2026-03-09 09:33:22 +0000 UTC" firstStartedPulling="2026-03-09 09:33:23.713034218 +0000 UTC m=+804.537769799" lastFinishedPulling="2026-03-09 09:33:28.004470441 +0000 UTC m=+808.829206022" observedRunningTime="2026-03-09 09:33:29.215289246 +0000 UTC m=+810.040024827" watchObservedRunningTime="2026-03-09 09:33:29.217787749 +0000 UTC m=+810.042523340" Mar 09 09:33:29 crc kubenswrapper[4692]: I0309 09:33:29.219936 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6nhdr" podStartSLOduration=4.288258342 podStartE2EDuration="7.219927201s" podCreationTimestamp="2026-03-09 09:33:22 +0000 UTC" firstStartedPulling="2026-03-09 09:33:25.065846021 +0000 UTC m=+805.890581602" lastFinishedPulling="2026-03-09 09:33:27.99751488 +0000 UTC m=+808.822250461" observedRunningTime="2026-03-09 09:33:29.17956555 +0000 UTC m=+810.004301131" watchObservedRunningTime="2026-03-09 09:33:29.219927201 +0000 UTC m=+810.044662782" Mar 09 09:33:33 crc kubenswrapper[4692]: I0309 09:33:33.206903 4692 generic.go:334] "Generic (PLEG): container finished" podID="32f4683a-ccb2-44de-a220-6e279f138d99" containerID="cec0999faca136833959cd29c09094ee2d9e67d00cf972a7eb3988886516174a" exitCode=0 Mar 09 09:33:33 crc kubenswrapper[4692]: I0309 09:33:33.207016 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerDied","Data":"cec0999faca136833959cd29c09094ee2d9e67d00cf972a7eb3988886516174a"} Mar 09 09:33:33 crc kubenswrapper[4692]: I0309 09:33:33.210773 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" event={"ID":"c1e68ce3-7af3-4509-a31e-81f42f223aa9","Type":"ContainerStarted","Data":"1f151b9239bc0978d6940169f495df2477a723e62dbd8ee16b7a6f4ba3b3dc82"} Mar 09 09:33:33 crc kubenswrapper[4692]: I0309 09:33:33.210952 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:33 crc kubenswrapper[4692]: I0309 09:33:33.250349 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" podStartSLOduration=2.009030386 podStartE2EDuration="11.250328631s" podCreationTimestamp="2026-03-09 09:33:22 +0000 UTC" firstStartedPulling="2026-03-09 09:33:23.469921183 +0000 UTC m=+804.294656764" lastFinishedPulling="2026-03-09 09:33:32.711219428 +0000 UTC m=+813.535955009" observedRunningTime="2026-03-09 09:33:33.248341754 +0000 UTC m=+814.073077365" watchObservedRunningTime="2026-03-09 09:33:33.250328631 +0000 UTC m=+814.075064212" Mar 09 09:33:33 crc kubenswrapper[4692]: I0309 09:33:33.329100 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-86ddb6bd46-fhbtn" Mar 09 09:33:34 crc kubenswrapper[4692]: I0309 09:33:34.216965 4692 generic.go:334] "Generic (PLEG): container finished" podID="32f4683a-ccb2-44de-a220-6e279f138d99" containerID="f4f172cd609d8a3424c6858ce4f7fcc3d9be750f1b95256de008b9ea121919b5" exitCode=0 Mar 09 09:33:34 crc kubenswrapper[4692]: I0309 09:33:34.217099 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerDied","Data":"f4f172cd609d8a3424c6858ce4f7fcc3d9be750f1b95256de008b9ea121919b5"} Mar 09 09:33:35 crc kubenswrapper[4692]: I0309 09:33:35.224266 4692 generic.go:334] "Generic (PLEG): container finished" podID="32f4683a-ccb2-44de-a220-6e279f138d99" containerID="10e212161a14a3572d28a2fb737aa1f18d7288ef4b20d6b1b7e8a5ae8cdb63f0" exitCode=0 Mar 09 09:33:35 crc kubenswrapper[4692]: I0309 09:33:35.224368 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerDied","Data":"10e212161a14a3572d28a2fb737aa1f18d7288ef4b20d6b1b7e8a5ae8cdb63f0"} Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.232219 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerStarted","Data":"cea26752d4299b78cfe65b47859dae2bd55afb273f9b40bfcb2809aab92d91af"} Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.232252 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerStarted","Data":"c13ce48e1fdd5ce612861359c2aa32a3ed18f4ce40556ec9c216f235da80648f"} Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.232263 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerStarted","Data":"9495cd128c4d8299d286f7fefe1eeca8f8defb1ddda6a4a10bfbc6b750f5281b"} Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.232271 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerStarted","Data":"145bd05a120388b857367504700b401b97442bd0bf3fc3895ac4088e079b73fa"} Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.232280 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerStarted","Data":"44ff73deec17af4f45fc83694deb5e4de67ec3067bf3e238820be540547f1ef1"} Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.232288 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-5b8n8" event={"ID":"32f4683a-ccb2-44de-a220-6e279f138d99","Type":"ContainerStarted","Data":"1aa94d90ef7771a47063db3f791e122dde773cdb3d33fe9354b887b0e99fc0c4"} Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.232370 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:36 crc kubenswrapper[4692]: I0309 09:33:36.270436 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-5b8n8" podStartSLOduration=5.511496589 podStartE2EDuration="14.270418806s" podCreationTimestamp="2026-03-09 09:33:22 +0000 UTC" firstStartedPulling="2026-03-09 09:33:23.976379529 +0000 UTC m=+804.801115110" lastFinishedPulling="2026-03-09 09:33:32.735301746 +0000 UTC m=+813.560037327" observedRunningTime="2026-03-09 09:33:36.270307153 +0000 UTC m=+817.095042744" watchObservedRunningTime="2026-03-09 09:33:36.270418806 +0000 UTC m=+817.095154397" Mar 09 09:33:38 crc kubenswrapper[4692]: I0309 09:33:38.816228 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:38 crc kubenswrapper[4692]: I0309 09:33:38.858542 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:43 crc kubenswrapper[4692]: I0309 09:33:43.228118 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-n6cmq" Mar 09 09:33:44 crc kubenswrapper[4692]: I0309 09:33:44.812630 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6nhdr" Mar 09 09:33:47 crc kubenswrapper[4692]: I0309 09:33:47.623836 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:33:47 crc kubenswrapper[4692]: I0309 09:33:47.624212 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:33:47 crc kubenswrapper[4692]: I0309 09:33:47.624264 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:33:47 crc kubenswrapper[4692]: I0309 09:33:47.624896 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fa3fd1dcbe9393b3fc6c43e0491320471ea475c9f05b37486ae66af62282e34"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:33:47 crc kubenswrapper[4692]: I0309 09:33:47.624968 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://5fa3fd1dcbe9393b3fc6c43e0491320471ea475c9f05b37486ae66af62282e34" gracePeriod=600 Mar 09 09:33:48 crc kubenswrapper[4692]: I0309 09:33:48.298595 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="5fa3fd1dcbe9393b3fc6c43e0491320471ea475c9f05b37486ae66af62282e34" exitCode=0 Mar 09 09:33:48 crc kubenswrapper[4692]: I0309 09:33:48.298640 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"5fa3fd1dcbe9393b3fc6c43e0491320471ea475c9f05b37486ae66af62282e34"} Mar 09 09:33:48 crc kubenswrapper[4692]: I0309 09:33:48.298664 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"c7bb4acc73ba36420c0329ef9cb241fb590012d44ec365a56134ad986b42e9b2"} Mar 09 09:33:48 crc kubenswrapper[4692]: I0309 09:33:48.298680 4692 scope.go:117] "RemoveContainer" containerID="31f9b48232438cf31701990fc67d0a37a49524e82d3bf2636f02423390a87487" Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.696191 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-df4cb"] Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.697638 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-df4cb" Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.699768 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.699899 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.704445 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-df4cb"] Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.708387 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-c6cwk" Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.851797 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9n42\" (UniqueName: \"kubernetes.io/projected/d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab-kube-api-access-c9n42\") pod \"mariadb-operator-index-df4cb\" (UID: \"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab\") " pod="openstack-operators/mariadb-operator-index-df4cb" Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.953429 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9n42\" (UniqueName: \"kubernetes.io/projected/d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab-kube-api-access-c9n42\") pod \"mariadb-operator-index-df4cb\" (UID: \"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab\") " pod="openstack-operators/mariadb-operator-index-df4cb" Mar 09 09:33:50 crc kubenswrapper[4692]: I0309 09:33:50.972279 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9n42\" (UniqueName: \"kubernetes.io/projected/d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab-kube-api-access-c9n42\") pod \"mariadb-operator-index-df4cb\" (UID: \"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab\") " pod="openstack-operators/mariadb-operator-index-df4cb" Mar 09 09:33:51 crc kubenswrapper[4692]: I0309 09:33:51.018507 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-df4cb" Mar 09 09:33:51 crc kubenswrapper[4692]: I0309 09:33:51.354298 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-df4cb"] Mar 09 09:33:52 crc kubenswrapper[4692]: I0309 09:33:52.349875 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-df4cb" event={"ID":"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab","Type":"ContainerStarted","Data":"ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0"} Mar 09 09:33:52 crc kubenswrapper[4692]: I0309 09:33:52.350252 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-df4cb" event={"ID":"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab","Type":"ContainerStarted","Data":"bcf1c4c7104fa0bcd80be7fdb0b9b3c105eba971470b8bd124867e88e7d7b4c4"} Mar 09 09:33:52 crc kubenswrapper[4692]: I0309 09:33:52.365444 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-df4cb" podStartSLOduration=1.627317699 podStartE2EDuration="2.365426627s" podCreationTimestamp="2026-03-09 09:33:50 +0000 UTC" firstStartedPulling="2026-03-09 09:33:51.371796304 +0000 UTC m=+832.196531885" lastFinishedPulling="2026-03-09 09:33:52.109905242 +0000 UTC m=+832.934640813" observedRunningTime="2026-03-09 09:33:52.360579316 +0000 UTC m=+833.185314917" watchObservedRunningTime="2026-03-09 09:33:52.365426627 +0000 UTC m=+833.190162208" Mar 09 09:33:53 crc kubenswrapper[4692]: I0309 09:33:53.818994 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-5b8n8" Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.081435 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-df4cb"] Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.364223 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-df4cb" podUID="d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab" containerName="registry-server" containerID="cri-o://ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0" gracePeriod=2 Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.689646 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-54n7l"] Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.690713 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.693098 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-54n7l"] Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.716634 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-df4cb" Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.801026 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9n42\" (UniqueName: \"kubernetes.io/projected/d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab-kube-api-access-c9n42\") pod \"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab\" (UID: \"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab\") " Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.801338 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbxwj\" (UniqueName: \"kubernetes.io/projected/8733e667-ec14-4dbd-919d-c69774fdfec0-kube-api-access-qbxwj\") pod \"mariadb-operator-index-54n7l\" (UID: \"8733e667-ec14-4dbd-919d-c69774fdfec0\") " pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.807400 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab-kube-api-access-c9n42" (OuterVolumeSpecName: "kube-api-access-c9n42") pod "d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab" (UID: "d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab"). InnerVolumeSpecName "kube-api-access-c9n42". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.902721 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbxwj\" (UniqueName: \"kubernetes.io/projected/8733e667-ec14-4dbd-919d-c69774fdfec0-kube-api-access-qbxwj\") pod \"mariadb-operator-index-54n7l\" (UID: \"8733e667-ec14-4dbd-919d-c69774fdfec0\") " pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.902839 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9n42\" (UniqueName: \"kubernetes.io/projected/d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab-kube-api-access-c9n42\") on node \"crc\" DevicePath \"\"" Mar 09 09:33:54 crc kubenswrapper[4692]: I0309 09:33:54.919112 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbxwj\" (UniqueName: \"kubernetes.io/projected/8733e667-ec14-4dbd-919d-c69774fdfec0-kube-api-access-qbxwj\") pod \"mariadb-operator-index-54n7l\" (UID: \"8733e667-ec14-4dbd-919d-c69774fdfec0\") " pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.030563 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.371724 4692 generic.go:334] "Generic (PLEG): container finished" podID="d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab" containerID="ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0" exitCode=0 Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.371791 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-df4cb" Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.371790 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-df4cb" event={"ID":"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab","Type":"ContainerDied","Data":"ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0"} Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.371866 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-df4cb" event={"ID":"d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab","Type":"ContainerDied","Data":"bcf1c4c7104fa0bcd80be7fdb0b9b3c105eba971470b8bd124867e88e7d7b4c4"} Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.371889 4692 scope.go:117] "RemoveContainer" containerID="ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0" Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.388417 4692 scope.go:117] "RemoveContainer" containerID="ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0" Mar 09 09:33:55 crc kubenswrapper[4692]: E0309 09:33:55.388972 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0\": container with ID starting with ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0 not found: ID does not exist" containerID="ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0" Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.389000 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0"} err="failed to get container status \"ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0\": rpc error: code = NotFound desc = could not find container \"ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0\": container with ID starting with ecd4111a95313001f02b8d966959235a7ac4f301b7b461bdacda1c17792316d0 not found: ID does not exist" Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.401121 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-df4cb"] Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.406462 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-df4cb"] Mar 09 09:33:55 crc kubenswrapper[4692]: I0309 09:33:55.417103 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-54n7l"] Mar 09 09:33:55 crc kubenswrapper[4692]: W0309 09:33:55.419236 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8733e667_ec14_4dbd_919d_c69774fdfec0.slice/crio-7ee879af55486a34425936841ee753b5de85491b0485b246cec19615fbfd0ffa WatchSource:0}: Error finding container 7ee879af55486a34425936841ee753b5de85491b0485b246cec19615fbfd0ffa: Status 404 returned error can't find the container with id 7ee879af55486a34425936841ee753b5de85491b0485b246cec19615fbfd0ffa Mar 09 09:33:56 crc kubenswrapper[4692]: I0309 09:33:56.081573 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab" path="/var/lib/kubelet/pods/d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab/volumes" Mar 09 09:33:56 crc kubenswrapper[4692]: I0309 09:33:56.379534 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-54n7l" event={"ID":"8733e667-ec14-4dbd-919d-c69774fdfec0","Type":"ContainerStarted","Data":"8d1e9361570b89adc46352faadc2d2568ceaa04a5de301d8ec5c249139935d72"} Mar 09 09:33:56 crc kubenswrapper[4692]: I0309 09:33:56.379583 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-54n7l" event={"ID":"8733e667-ec14-4dbd-919d-c69774fdfec0","Type":"ContainerStarted","Data":"7ee879af55486a34425936841ee753b5de85491b0485b246cec19615fbfd0ffa"} Mar 09 09:33:56 crc kubenswrapper[4692]: I0309 09:33:56.395004 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-54n7l" podStartSLOduration=1.849500184 podStartE2EDuration="2.394980322s" podCreationTimestamp="2026-03-09 09:33:54 +0000 UTC" firstStartedPulling="2026-03-09 09:33:55.422586967 +0000 UTC m=+836.247322548" lastFinishedPulling="2026-03-09 09:33:55.968067085 +0000 UTC m=+836.792802686" observedRunningTime="2026-03-09 09:33:56.392763247 +0000 UTC m=+837.217498848" watchObservedRunningTime="2026-03-09 09:33:56.394980322 +0000 UTC m=+837.219715903" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.125908 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550814-wj7qv"] Mar 09 09:34:00 crc kubenswrapper[4692]: E0309 09:34:00.126783 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab" containerName="registry-server" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.126799 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab" containerName="registry-server" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.126937 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2af5f16-2b2e-4dee-b06f-ccbb0cb499ab" containerName="registry-server" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.127380 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550814-wj7qv" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.132341 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.132424 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.135742 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.136774 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550814-wj7qv"] Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.265958 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbtpv\" (UniqueName: \"kubernetes.io/projected/f9c73711-7e88-4f9a-a2fc-3ace9e4144ec-kube-api-access-nbtpv\") pod \"auto-csr-approver-29550814-wj7qv\" (UID: \"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec\") " pod="openshift-infra/auto-csr-approver-29550814-wj7qv" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.367686 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbtpv\" (UniqueName: \"kubernetes.io/projected/f9c73711-7e88-4f9a-a2fc-3ace9e4144ec-kube-api-access-nbtpv\") pod \"auto-csr-approver-29550814-wj7qv\" (UID: \"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec\") " pod="openshift-infra/auto-csr-approver-29550814-wj7qv" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.390333 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbtpv\" (UniqueName: \"kubernetes.io/projected/f9c73711-7e88-4f9a-a2fc-3ace9e4144ec-kube-api-access-nbtpv\") pod \"auto-csr-approver-29550814-wj7qv\" (UID: \"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec\") " pod="openshift-infra/auto-csr-approver-29550814-wj7qv" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.445664 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550814-wj7qv" Mar 09 09:34:00 crc kubenswrapper[4692]: I0309 09:34:00.629338 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550814-wj7qv"] Mar 09 09:34:00 crc kubenswrapper[4692]: W0309 09:34:00.637384 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9c73711_7e88_4f9a_a2fc_3ace9e4144ec.slice/crio-5f3bfc1b9ddd062d8d1a7d9aa709648d72b02836c392c403c8338a455a9d21e0 WatchSource:0}: Error finding container 5f3bfc1b9ddd062d8d1a7d9aa709648d72b02836c392c403c8338a455a9d21e0: Status 404 returned error can't find the container with id 5f3bfc1b9ddd062d8d1a7d9aa709648d72b02836c392c403c8338a455a9d21e0 Mar 09 09:34:01 crc kubenswrapper[4692]: I0309 09:34:01.406787 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550814-wj7qv" event={"ID":"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec","Type":"ContainerStarted","Data":"5f3bfc1b9ddd062d8d1a7d9aa709648d72b02836c392c403c8338a455a9d21e0"} Mar 09 09:34:02 crc kubenswrapper[4692]: I0309 09:34:02.414507 4692 generic.go:334] "Generic (PLEG): container finished" podID="f9c73711-7e88-4f9a-a2fc-3ace9e4144ec" containerID="c47374ee0cef70824c305d4e806b22e9b170ce84fa698597c397a090752e0f3e" exitCode=0 Mar 09 09:34:02 crc kubenswrapper[4692]: I0309 09:34:02.414852 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550814-wj7qv" event={"ID":"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec","Type":"ContainerDied","Data":"c47374ee0cef70824c305d4e806b22e9b170ce84fa698597c397a090752e0f3e"} Mar 09 09:34:03 crc kubenswrapper[4692]: I0309 09:34:03.657944 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550814-wj7qv" Mar 09 09:34:03 crc kubenswrapper[4692]: I0309 09:34:03.809545 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbtpv\" (UniqueName: \"kubernetes.io/projected/f9c73711-7e88-4f9a-a2fc-3ace9e4144ec-kube-api-access-nbtpv\") pod \"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec\" (UID: \"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec\") " Mar 09 09:34:03 crc kubenswrapper[4692]: I0309 09:34:03.818443 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9c73711-7e88-4f9a-a2fc-3ace9e4144ec-kube-api-access-nbtpv" (OuterVolumeSpecName: "kube-api-access-nbtpv") pod "f9c73711-7e88-4f9a-a2fc-3ace9e4144ec" (UID: "f9c73711-7e88-4f9a-a2fc-3ace9e4144ec"). InnerVolumeSpecName "kube-api-access-nbtpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:34:03 crc kubenswrapper[4692]: I0309 09:34:03.911625 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbtpv\" (UniqueName: \"kubernetes.io/projected/f9c73711-7e88-4f9a-a2fc-3ace9e4144ec-kube-api-access-nbtpv\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:04 crc kubenswrapper[4692]: I0309 09:34:04.431360 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550814-wj7qv" event={"ID":"f9c73711-7e88-4f9a-a2fc-3ace9e4144ec","Type":"ContainerDied","Data":"5f3bfc1b9ddd062d8d1a7d9aa709648d72b02836c392c403c8338a455a9d21e0"} Mar 09 09:34:04 crc kubenswrapper[4692]: I0309 09:34:04.431399 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f3bfc1b9ddd062d8d1a7d9aa709648d72b02836c392c403c8338a455a9d21e0" Mar 09 09:34:04 crc kubenswrapper[4692]: I0309 09:34:04.431422 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550814-wj7qv" Mar 09 09:34:04 crc kubenswrapper[4692]: I0309 09:34:04.706087 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550808-shsh8"] Mar 09 09:34:04 crc kubenswrapper[4692]: I0309 09:34:04.709573 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550808-shsh8"] Mar 09 09:34:05 crc kubenswrapper[4692]: I0309 09:34:05.031153 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:34:05 crc kubenswrapper[4692]: I0309 09:34:05.031226 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:34:05 crc kubenswrapper[4692]: I0309 09:34:05.059669 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:34:05 crc kubenswrapper[4692]: I0309 09:34:05.467375 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-54n7l" Mar 09 09:34:06 crc kubenswrapper[4692]: I0309 09:34:06.078851 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="196f410f-b0a7-4bac-851f-2910d63c5b0f" path="/var/lib/kubelet/pods/196f410f-b0a7-4bac-851f-2910d63c5b0f/volumes" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.825633 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw"] Mar 09 09:34:10 crc kubenswrapper[4692]: E0309 09:34:10.826339 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c73711-7e88-4f9a-a2fc-3ace9e4144ec" containerName="oc" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.826351 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c73711-7e88-4f9a-a2fc-3ace9e4144ec" containerName="oc" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.826452 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c73711-7e88-4f9a-a2fc-3ace9e4144ec" containerName="oc" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.827277 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.829495 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lthqw" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.836500 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw"] Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.995252 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.995300 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skppm\" (UniqueName: \"kubernetes.io/projected/09cae4f4-3bef-442c-88ad-a84932acf202-kube-api-access-skppm\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:10 crc kubenswrapper[4692]: I0309 09:34:10.995375 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.097010 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.097070 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.097093 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skppm\" (UniqueName: \"kubernetes.io/projected/09cae4f4-3bef-442c-88ad-a84932acf202-kube-api-access-skppm\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.097493 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.097565 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.118128 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skppm\" (UniqueName: \"kubernetes.io/projected/09cae4f4-3bef-442c-88ad-a84932acf202-kube-api-access-skppm\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.146348 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.328521 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw"] Mar 09 09:34:11 crc kubenswrapper[4692]: I0309 09:34:11.469635 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" event={"ID":"09cae4f4-3bef-442c-88ad-a84932acf202","Type":"ContainerStarted","Data":"f3c58fe17fb4b2d5500eb2aabf3714be01239a90633c50789a5eede3fa8ed628"} Mar 09 09:34:12 crc kubenswrapper[4692]: I0309 09:34:12.477393 4692 generic.go:334] "Generic (PLEG): container finished" podID="09cae4f4-3bef-442c-88ad-a84932acf202" containerID="f94e00282e09a85b3283e39caba019ccdbbea6a4f2ee6e9f62711a5fdb9047ee" exitCode=0 Mar 09 09:34:12 crc kubenswrapper[4692]: I0309 09:34:12.477433 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" event={"ID":"09cae4f4-3bef-442c-88ad-a84932acf202","Type":"ContainerDied","Data":"f94e00282e09a85b3283e39caba019ccdbbea6a4f2ee6e9f62711a5fdb9047ee"} Mar 09 09:34:13 crc kubenswrapper[4692]: I0309 09:34:13.484422 4692 generic.go:334] "Generic (PLEG): container finished" podID="09cae4f4-3bef-442c-88ad-a84932acf202" containerID="36a2d0750cfc15ee713a0d4d72dd47b52501b22aa2955b09863c5b083c7011af" exitCode=0 Mar 09 09:34:13 crc kubenswrapper[4692]: I0309 09:34:13.484469 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" event={"ID":"09cae4f4-3bef-442c-88ad-a84932acf202","Type":"ContainerDied","Data":"36a2d0750cfc15ee713a0d4d72dd47b52501b22aa2955b09863c5b083c7011af"} Mar 09 09:34:14 crc kubenswrapper[4692]: I0309 09:34:14.491545 4692 generic.go:334] "Generic (PLEG): container finished" podID="09cae4f4-3bef-442c-88ad-a84932acf202" containerID="fa43fca225ea818d0dd867c04a897771abd6c9b1cd7a0be8ee74625876a77707" exitCode=0 Mar 09 09:34:14 crc kubenswrapper[4692]: I0309 09:34:14.491643 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" event={"ID":"09cae4f4-3bef-442c-88ad-a84932acf202","Type":"ContainerDied","Data":"fa43fca225ea818d0dd867c04a897771abd6c9b1cd7a0be8ee74625876a77707"} Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.719936 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.855885 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-util\") pod \"09cae4f4-3bef-442c-88ad-a84932acf202\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.855964 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skppm\" (UniqueName: \"kubernetes.io/projected/09cae4f4-3bef-442c-88ad-a84932acf202-kube-api-access-skppm\") pod \"09cae4f4-3bef-442c-88ad-a84932acf202\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.856040 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-bundle\") pod \"09cae4f4-3bef-442c-88ad-a84932acf202\" (UID: \"09cae4f4-3bef-442c-88ad-a84932acf202\") " Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.857245 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-bundle" (OuterVolumeSpecName: "bundle") pod "09cae4f4-3bef-442c-88ad-a84932acf202" (UID: "09cae4f4-3bef-442c-88ad-a84932acf202"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.861340 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09cae4f4-3bef-442c-88ad-a84932acf202-kube-api-access-skppm" (OuterVolumeSpecName: "kube-api-access-skppm") pod "09cae4f4-3bef-442c-88ad-a84932acf202" (UID: "09cae4f4-3bef-442c-88ad-a84932acf202"). InnerVolumeSpecName "kube-api-access-skppm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.871722 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-util" (OuterVolumeSpecName: "util") pod "09cae4f4-3bef-442c-88ad-a84932acf202" (UID: "09cae4f4-3bef-442c-88ad-a84932acf202"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.957084 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skppm\" (UniqueName: \"kubernetes.io/projected/09cae4f4-3bef-442c-88ad-a84932acf202-kube-api-access-skppm\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.957122 4692 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:15 crc kubenswrapper[4692]: I0309 09:34:15.957130 4692 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/09cae4f4-3bef-442c-88ad-a84932acf202-util\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:16 crc kubenswrapper[4692]: I0309 09:34:16.503088 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" event={"ID":"09cae4f4-3bef-442c-88ad-a84932acf202","Type":"ContainerDied","Data":"f3c58fe17fb4b2d5500eb2aabf3714be01239a90633c50789a5eede3fa8ed628"} Mar 09 09:34:16 crc kubenswrapper[4692]: I0309 09:34:16.503431 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3c58fe17fb4b2d5500eb2aabf3714be01239a90633c50789a5eede3fa8ed628" Mar 09 09:34:16 crc kubenswrapper[4692]: I0309 09:34:16.503235 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.760783 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct"] Mar 09 09:34:23 crc kubenswrapper[4692]: E0309 09:34:23.761535 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cae4f4-3bef-442c-88ad-a84932acf202" containerName="pull" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.761547 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cae4f4-3bef-442c-88ad-a84932acf202" containerName="pull" Mar 09 09:34:23 crc kubenswrapper[4692]: E0309 09:34:23.761562 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cae4f4-3bef-442c-88ad-a84932acf202" containerName="util" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.761568 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cae4f4-3bef-442c-88ad-a84932acf202" containerName="util" Mar 09 09:34:23 crc kubenswrapper[4692]: E0309 09:34:23.761576 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cae4f4-3bef-442c-88ad-a84932acf202" containerName="extract" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.761582 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cae4f4-3bef-442c-88ad-a84932acf202" containerName="extract" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.761669 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="09cae4f4-3bef-442c-88ad-a84932acf202" containerName="extract" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.762066 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.765919 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.766372 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.767228 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-55nkg" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.797341 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct"] Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.848218 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lz8l\" (UniqueName: \"kubernetes.io/projected/3336031e-a09e-471f-858c-f9c66f040451-kube-api-access-2lz8l\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.848328 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3336031e-a09e-471f-858c-f9c66f040451-apiservice-cert\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.848357 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3336031e-a09e-471f-858c-f9c66f040451-webhook-cert\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.949465 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lz8l\" (UniqueName: \"kubernetes.io/projected/3336031e-a09e-471f-858c-f9c66f040451-kube-api-access-2lz8l\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.949579 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3336031e-a09e-471f-858c-f9c66f040451-apiservice-cert\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.949607 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3336031e-a09e-471f-858c-f9c66f040451-webhook-cert\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.955438 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3336031e-a09e-471f-858c-f9c66f040451-apiservice-cert\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.955966 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3336031e-a09e-471f-858c-f9c66f040451-webhook-cert\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:23 crc kubenswrapper[4692]: I0309 09:34:23.967123 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lz8l\" (UniqueName: \"kubernetes.io/projected/3336031e-a09e-471f-858c-f9c66f040451-kube-api-access-2lz8l\") pod \"mariadb-operator-controller-manager-b4fb85cb4-f9kct\" (UID: \"3336031e-a09e-471f-858c-f9c66f040451\") " pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:24 crc kubenswrapper[4692]: I0309 09:34:24.082973 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:24 crc kubenswrapper[4692]: I0309 09:34:24.276989 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct"] Mar 09 09:34:24 crc kubenswrapper[4692]: I0309 09:34:24.545762 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" event={"ID":"3336031e-a09e-471f-858c-f9c66f040451","Type":"ContainerStarted","Data":"a37094ed3d4cea7fffa12c804eaa9b0fd5bd23a3d99dcccd8710fede4338f664"} Mar 09 09:34:28 crc kubenswrapper[4692]: I0309 09:34:28.567435 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" event={"ID":"3336031e-a09e-471f-858c-f9c66f040451","Type":"ContainerStarted","Data":"7e8921501fe5a64d8627c858f4d3f44c7890fbe2b6a54148ece434c6f711f8ec"} Mar 09 09:34:28 crc kubenswrapper[4692]: I0309 09:34:28.568937 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:28 crc kubenswrapper[4692]: I0309 09:34:28.582213 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" podStartSLOduration=2.156045974 podStartE2EDuration="5.58219539s" podCreationTimestamp="2026-03-09 09:34:23 +0000 UTC" firstStartedPulling="2026-03-09 09:34:24.289311865 +0000 UTC m=+865.114047446" lastFinishedPulling="2026-03-09 09:34:27.715461281 +0000 UTC m=+868.540196862" observedRunningTime="2026-03-09 09:34:28.581422398 +0000 UTC m=+869.406157989" watchObservedRunningTime="2026-03-09 09:34:28.58219539 +0000 UTC m=+869.406930971" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.308091 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wp4gf"] Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.310726 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.320998 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wp4gf"] Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.359063 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c2af711-e30f-4153-85b8-864a5250e4e2-catalog-content\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.359112 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c2af711-e30f-4153-85b8-864a5250e4e2-utilities\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.359139 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snqtr\" (UniqueName: \"kubernetes.io/projected/2c2af711-e30f-4153-85b8-864a5250e4e2-kube-api-access-snqtr\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.460057 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c2af711-e30f-4153-85b8-864a5250e4e2-catalog-content\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.460116 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c2af711-e30f-4153-85b8-864a5250e4e2-utilities\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.460150 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snqtr\" (UniqueName: \"kubernetes.io/projected/2c2af711-e30f-4153-85b8-864a5250e4e2-kube-api-access-snqtr\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.461006 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c2af711-e30f-4153-85b8-864a5250e4e2-catalog-content\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.461299 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c2af711-e30f-4153-85b8-864a5250e4e2-utilities\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.494186 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snqtr\" (UniqueName: \"kubernetes.io/projected/2c2af711-e30f-4153-85b8-864a5250e4e2-kube-api-access-snqtr\") pod \"community-operators-wp4gf\" (UID: \"2c2af711-e30f-4153-85b8-864a5250e4e2\") " pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:32 crc kubenswrapper[4692]: I0309 09:34:32.637848 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:33 crc kubenswrapper[4692]: I0309 09:34:33.028289 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wp4gf"] Mar 09 09:34:33 crc kubenswrapper[4692]: I0309 09:34:33.594033 4692 generic.go:334] "Generic (PLEG): container finished" podID="2c2af711-e30f-4153-85b8-864a5250e4e2" containerID="e58a9b348ddd702400fe49d94bd8754e22e53c468cbf736d0400d528b361dcad" exitCode=0 Mar 09 09:34:33 crc kubenswrapper[4692]: I0309 09:34:33.594096 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4gf" event={"ID":"2c2af711-e30f-4153-85b8-864a5250e4e2","Type":"ContainerDied","Data":"e58a9b348ddd702400fe49d94bd8754e22e53c468cbf736d0400d528b361dcad"} Mar 09 09:34:33 crc kubenswrapper[4692]: I0309 09:34:33.594304 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4gf" event={"ID":"2c2af711-e30f-4153-85b8-864a5250e4e2","Type":"ContainerStarted","Data":"94abc765e11014d7cdd9f2539896f6a5297365b90645ace31b42deaae29dd109"} Mar 09 09:34:34 crc kubenswrapper[4692]: I0309 09:34:34.087703 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-b4fb85cb4-f9kct" Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.294306 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-ngw2q"] Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.295428 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.297624 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-jtwm5" Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.307236 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-ngw2q"] Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.408443 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wj2sd\" (UniqueName: \"kubernetes.io/projected/a26e421d-abf7-477e-bad5-6ce43d8dd862-kube-api-access-wj2sd\") pod \"infra-operator-index-ngw2q\" (UID: \"a26e421d-abf7-477e-bad5-6ce43d8dd862\") " pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.509179 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wj2sd\" (UniqueName: \"kubernetes.io/projected/a26e421d-abf7-477e-bad5-6ce43d8dd862-kube-api-access-wj2sd\") pod \"infra-operator-index-ngw2q\" (UID: \"a26e421d-abf7-477e-bad5-6ce43d8dd862\") " pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.562559 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wj2sd\" (UniqueName: \"kubernetes.io/projected/a26e421d-abf7-477e-bad5-6ce43d8dd862-kube-api-access-wj2sd\") pod \"infra-operator-index-ngw2q\" (UID: \"a26e421d-abf7-477e-bad5-6ce43d8dd862\") " pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.610899 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:40 crc kubenswrapper[4692]: I0309 09:34:40.699908 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4gf" event={"ID":"2c2af711-e30f-4153-85b8-864a5250e4e2","Type":"ContainerStarted","Data":"769cc1fb726472e46b4e7703e015c0b31216f74c70e385f8be8061d6653f1138"} Mar 09 09:34:41 crc kubenswrapper[4692]: I0309 09:34:41.088749 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-ngw2q"] Mar 09 09:34:41 crc kubenswrapper[4692]: I0309 09:34:41.706056 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ngw2q" event={"ID":"a26e421d-abf7-477e-bad5-6ce43d8dd862","Type":"ContainerStarted","Data":"7a1186889b672f98da8bba023ea89897e66da6105b51f55dffd796e0c8a3e0e7"} Mar 09 09:34:41 crc kubenswrapper[4692]: I0309 09:34:41.708780 4692 generic.go:334] "Generic (PLEG): container finished" podID="2c2af711-e30f-4153-85b8-864a5250e4e2" containerID="769cc1fb726472e46b4e7703e015c0b31216f74c70e385f8be8061d6653f1138" exitCode=0 Mar 09 09:34:41 crc kubenswrapper[4692]: I0309 09:34:41.708827 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4gf" event={"ID":"2c2af711-e30f-4153-85b8-864a5250e4e2","Type":"ContainerDied","Data":"769cc1fb726472e46b4e7703e015c0b31216f74c70e385f8be8061d6653f1138"} Mar 09 09:34:42 crc kubenswrapper[4692]: I0309 09:34:42.716956 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-ngw2q" event={"ID":"a26e421d-abf7-477e-bad5-6ce43d8dd862","Type":"ContainerStarted","Data":"09f8a70b692569aae2863d72fa7cf851922b9bf8c37fa9af270191230c37aaef"} Mar 09 09:34:42 crc kubenswrapper[4692]: I0309 09:34:42.719551 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wp4gf" event={"ID":"2c2af711-e30f-4153-85b8-864a5250e4e2","Type":"ContainerStarted","Data":"0aab7fa80ff449d36af1fd3ef0db290647eacd9d537f319e417299061ceda37d"} Mar 09 09:34:42 crc kubenswrapper[4692]: I0309 09:34:42.736115 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-ngw2q" podStartSLOduration=2.072404217 podStartE2EDuration="2.736096245s" podCreationTimestamp="2026-03-09 09:34:40 +0000 UTC" firstStartedPulling="2026-03-09 09:34:41.099769184 +0000 UTC m=+881.924504765" lastFinishedPulling="2026-03-09 09:34:41.763461212 +0000 UTC m=+882.588196793" observedRunningTime="2026-03-09 09:34:42.734628173 +0000 UTC m=+883.559363754" watchObservedRunningTime="2026-03-09 09:34:42.736096245 +0000 UTC m=+883.560831826" Mar 09 09:34:42 crc kubenswrapper[4692]: I0309 09:34:42.764973 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wp4gf" podStartSLOduration=2.234579546 podStartE2EDuration="10.764955053s" podCreationTimestamp="2026-03-09 09:34:32 +0000 UTC" firstStartedPulling="2026-03-09 09:34:33.595407689 +0000 UTC m=+874.420143260" lastFinishedPulling="2026-03-09 09:34:42.125783186 +0000 UTC m=+882.950518767" observedRunningTime="2026-03-09 09:34:42.761324327 +0000 UTC m=+883.586059908" watchObservedRunningTime="2026-03-09 09:34:42.764955053 +0000 UTC m=+883.589690644" Mar 09 09:34:50 crc kubenswrapper[4692]: I0309 09:34:50.611813 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:50 crc kubenswrapper[4692]: I0309 09:34:50.612464 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:50 crc kubenswrapper[4692]: I0309 09:34:50.678043 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:50 crc kubenswrapper[4692]: I0309 09:34:50.794964 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-ngw2q" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.119147 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt"] Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.120325 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.123326 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lthqw" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.139611 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt"] Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.172306 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmlds\" (UniqueName: \"kubernetes.io/projected/381ff66a-6f6e-49c5-9c29-d414b7200dae-kube-api-access-zmlds\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.172369 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-bundle\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.172408 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-util\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.273958 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmlds\" (UniqueName: \"kubernetes.io/projected/381ff66a-6f6e-49c5-9c29-d414b7200dae-kube-api-access-zmlds\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.274025 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-bundle\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.274069 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-util\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.274594 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-bundle\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.274646 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-util\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.299975 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmlds\" (UniqueName: \"kubernetes.io/projected/381ff66a-6f6e-49c5-9c29-d414b7200dae-kube-api-access-zmlds\") pod \"c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.437386 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.639421 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.639475 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.701960 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:52 crc kubenswrapper[4692]: I0309 09:34:52.833584 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wp4gf" Mar 09 09:34:53 crc kubenswrapper[4692]: I0309 09:34:53.192403 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt"] Mar 09 09:34:53 crc kubenswrapper[4692]: W0309 09:34:53.197426 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod381ff66a_6f6e_49c5_9c29_d414b7200dae.slice/crio-a2695cf77d30962d2934d8b442ce3c02dad043b4eef38e49882a8bb024211cd9 WatchSource:0}: Error finding container a2695cf77d30962d2934d8b442ce3c02dad043b4eef38e49882a8bb024211cd9: Status 404 returned error can't find the container with id a2695cf77d30962d2934d8b442ce3c02dad043b4eef38e49882a8bb024211cd9 Mar 09 09:34:53 crc kubenswrapper[4692]: I0309 09:34:53.784653 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" event={"ID":"381ff66a-6f6e-49c5-9c29-d414b7200dae","Type":"ContainerStarted","Data":"a2695cf77d30962d2934d8b442ce3c02dad043b4eef38e49882a8bb024211cd9"} Mar 09 09:34:54 crc kubenswrapper[4692]: I0309 09:34:54.790772 4692 generic.go:334] "Generic (PLEG): container finished" podID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerID="3a0dfc233e157a57b6d6ff2e6f6ea4602d9a84a8d9baa1b9aaec55d71d82a6bc" exitCode=0 Mar 09 09:34:54 crc kubenswrapper[4692]: I0309 09:34:54.790819 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" event={"ID":"381ff66a-6f6e-49c5-9c29-d414b7200dae","Type":"ContainerDied","Data":"3a0dfc233e157a57b6d6ff2e6f6ea4602d9a84a8d9baa1b9aaec55d71d82a6bc"} Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.694433 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6zl49"] Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.695531 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.710300 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zl49"] Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.799021 4692 generic.go:334] "Generic (PLEG): container finished" podID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerID="8e7028c56811027e64abca4e3425ca610f99f718fdda4884844dbb51d43c843c" exitCode=0 Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.799079 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" event={"ID":"381ff66a-6f6e-49c5-9c29-d414b7200dae","Type":"ContainerDied","Data":"8e7028c56811027e64abca4e3425ca610f99f718fdda4884844dbb51d43c843c"} Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.819834 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-utilities\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.819951 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tnnl\" (UniqueName: \"kubernetes.io/projected/f5a30e59-e0e0-4a43-be50-474183fb6db9-kube-api-access-5tnnl\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.819996 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-catalog-content\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.921257 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-utilities\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.921317 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tnnl\" (UniqueName: \"kubernetes.io/projected/f5a30e59-e0e0-4a43-be50-474183fb6db9-kube-api-access-5tnnl\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.921339 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-catalog-content\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.921947 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-utilities\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.922408 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-catalog-content\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:55 crc kubenswrapper[4692]: I0309 09:34:55.953324 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tnnl\" (UniqueName: \"kubernetes.io/projected/f5a30e59-e0e0-4a43-be50-474183fb6db9-kube-api-access-5tnnl\") pod \"redhat-marketplace-6zl49\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:56 crc kubenswrapper[4692]: I0309 09:34:56.029976 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:34:56 crc kubenswrapper[4692]: I0309 09:34:56.247784 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zl49"] Mar 09 09:34:56 crc kubenswrapper[4692]: I0309 09:34:56.806997 4692 generic.go:334] "Generic (PLEG): container finished" podID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerID="95c4df9c258b54951e32db42fa1ed97736a05ad565bcaddead407bd68c87061e" exitCode=0 Mar 09 09:34:56 crc kubenswrapper[4692]: I0309 09:34:56.807076 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" event={"ID":"381ff66a-6f6e-49c5-9c29-d414b7200dae","Type":"ContainerDied","Data":"95c4df9c258b54951e32db42fa1ed97736a05ad565bcaddead407bd68c87061e"} Mar 09 09:34:56 crc kubenswrapper[4692]: I0309 09:34:56.808805 4692 generic.go:334] "Generic (PLEG): container finished" podID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerID="dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62" exitCode=0 Mar 09 09:34:56 crc kubenswrapper[4692]: I0309 09:34:56.808851 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zl49" event={"ID":"f5a30e59-e0e0-4a43-be50-474183fb6db9","Type":"ContainerDied","Data":"dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62"} Mar 09 09:34:56 crc kubenswrapper[4692]: I0309 09:34:56.808881 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zl49" event={"ID":"f5a30e59-e0e0-4a43-be50-474183fb6db9","Type":"ContainerStarted","Data":"8eeb52651b16e7cb5bcc92bc38003bb5ce2fb358cac5378c2a8b50a9db27cc7e"} Mar 09 09:34:57 crc kubenswrapper[4692]: I0309 09:34:57.814719 4692 generic.go:334] "Generic (PLEG): container finished" podID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerID="63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893" exitCode=0 Mar 09 09:34:57 crc kubenswrapper[4692]: I0309 09:34:57.814926 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zl49" event={"ID":"f5a30e59-e0e0-4a43-be50-474183fb6db9","Type":"ContainerDied","Data":"63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893"} Mar 09 09:34:57 crc kubenswrapper[4692]: I0309 09:34:57.909060 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wp4gf"] Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.156362 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.251500 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-bundle\") pod \"381ff66a-6f6e-49c5-9c29-d414b7200dae\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.251558 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-util\") pod \"381ff66a-6f6e-49c5-9c29-d414b7200dae\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.251657 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmlds\" (UniqueName: \"kubernetes.io/projected/381ff66a-6f6e-49c5-9c29-d414b7200dae-kube-api-access-zmlds\") pod \"381ff66a-6f6e-49c5-9c29-d414b7200dae\" (UID: \"381ff66a-6f6e-49c5-9c29-d414b7200dae\") " Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.253811 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-bundle" (OuterVolumeSpecName: "bundle") pod "381ff66a-6f6e-49c5-9c29-d414b7200dae" (UID: "381ff66a-6f6e-49c5-9c29-d414b7200dae"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.268220 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-util" (OuterVolumeSpecName: "util") pod "381ff66a-6f6e-49c5-9c29-d414b7200dae" (UID: "381ff66a-6f6e-49c5-9c29-d414b7200dae"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.269754 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/381ff66a-6f6e-49c5-9c29-d414b7200dae-kube-api-access-zmlds" (OuterVolumeSpecName: "kube-api-access-zmlds") pod "381ff66a-6f6e-49c5-9c29-d414b7200dae" (UID: "381ff66a-6f6e-49c5-9c29-d414b7200dae"). InnerVolumeSpecName "kube-api-access-zmlds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.352830 4692 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.352858 4692 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/381ff66a-6f6e-49c5-9c29-d414b7200dae-util\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.352871 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmlds\" (UniqueName: \"kubernetes.io/projected/381ff66a-6f6e-49c5-9c29-d414b7200dae-kube-api-access-zmlds\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.675501 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-68gtj"] Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.675786 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-68gtj" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="registry-server" containerID="cri-o://522041d8e3397f7b3009880124ed49c62f56b5c13c5e15a8881aa7b83a50b75e" gracePeriod=2 Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.826902 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zl49" event={"ID":"f5a30e59-e0e0-4a43-be50-474183fb6db9","Type":"ContainerStarted","Data":"065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778"} Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.830137 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" event={"ID":"381ff66a-6f6e-49c5-9c29-d414b7200dae","Type":"ContainerDied","Data":"a2695cf77d30962d2934d8b442ce3c02dad043b4eef38e49882a8bb024211cd9"} Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.830446 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2695cf77d30962d2934d8b442ce3c02dad043b4eef38e49882a8bb024211cd9" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.830190 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt" Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.835199 4692 generic.go:334] "Generic (PLEG): container finished" podID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerID="522041d8e3397f7b3009880124ed49c62f56b5c13c5e15a8881aa7b83a50b75e" exitCode=0 Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.835263 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68gtj" event={"ID":"1df3c8b1-25ba-4b83-abdd-bf70af81208c","Type":"ContainerDied","Data":"522041d8e3397f7b3009880124ed49c62f56b5c13c5e15a8881aa7b83a50b75e"} Mar 09 09:34:58 crc kubenswrapper[4692]: I0309 09:34:58.862009 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6zl49" podStartSLOduration=2.2410039680000002 podStartE2EDuration="3.861988774s" podCreationTimestamp="2026-03-09 09:34:55 +0000 UTC" firstStartedPulling="2026-03-09 09:34:56.810545868 +0000 UTC m=+897.635281449" lastFinishedPulling="2026-03-09 09:34:58.431530674 +0000 UTC m=+899.256266255" observedRunningTime="2026-03-09 09:34:58.857072342 +0000 UTC m=+899.681807933" watchObservedRunningTime="2026-03-09 09:34:58.861988774 +0000 UTC m=+899.686724365" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.051975 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.161940 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-catalog-content\") pod \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.162009 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n4b5\" (UniqueName: \"kubernetes.io/projected/1df3c8b1-25ba-4b83-abdd-bf70af81208c-kube-api-access-4n4b5\") pod \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.162063 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-utilities\") pod \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\" (UID: \"1df3c8b1-25ba-4b83-abdd-bf70af81208c\") " Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.162984 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-utilities" (OuterVolumeSpecName: "utilities") pod "1df3c8b1-25ba-4b83-abdd-bf70af81208c" (UID: "1df3c8b1-25ba-4b83-abdd-bf70af81208c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.172119 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1df3c8b1-25ba-4b83-abdd-bf70af81208c-kube-api-access-4n4b5" (OuterVolumeSpecName: "kube-api-access-4n4b5") pod "1df3c8b1-25ba-4b83-abdd-bf70af81208c" (UID: "1df3c8b1-25ba-4b83-abdd-bf70af81208c"). InnerVolumeSpecName "kube-api-access-4n4b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.227105 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1df3c8b1-25ba-4b83-abdd-bf70af81208c" (UID: "1df3c8b1-25ba-4b83-abdd-bf70af81208c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.263773 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.264035 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1df3c8b1-25ba-4b83-abdd-bf70af81208c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.264101 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n4b5\" (UniqueName: \"kubernetes.io/projected/1df3c8b1-25ba-4b83-abdd-bf70af81208c-kube-api-access-4n4b5\") on node \"crc\" DevicePath \"\"" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.843939 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68gtj" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.843935 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68gtj" event={"ID":"1df3c8b1-25ba-4b83-abdd-bf70af81208c","Type":"ContainerDied","Data":"30aa005897999dcd1f7e231530f8409996567fc2b305ec9ae23cbda04825498d"} Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.844080 4692 scope.go:117] "RemoveContainer" containerID="522041d8e3397f7b3009880124ed49c62f56b5c13c5e15a8881aa7b83a50b75e" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.862033 4692 scope.go:117] "RemoveContainer" containerID="d575fc5324e973ef932620dae0ac1c8d6d04170c3dcc03813ebfd838ac1618b4" Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.878806 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-68gtj"] Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.879516 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-68gtj"] Mar 09 09:34:59 crc kubenswrapper[4692]: I0309 09:34:59.885681 4692 scope.go:117] "RemoveContainer" containerID="c2217d668c24d290125792a77e8944b4ebee1c33814afbb22098ae14a53819fe" Mar 09 09:35:00 crc kubenswrapper[4692]: I0309 09:35:00.078774 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" path="/var/lib/kubelet/pods/1df3c8b1-25ba-4b83-abdd-bf70af81208c/volumes" Mar 09 09:35:01 crc kubenswrapper[4692]: I0309 09:35:01.048134 4692 scope.go:117] "RemoveContainer" containerID="530727072ce5566044dbb1c8874daac2dc304d4e46a651b7cf913f6748ed3967" Mar 09 09:35:06 crc kubenswrapper[4692]: I0309 09:35:06.030709 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:35:06 crc kubenswrapper[4692]: I0309 09:35:06.031084 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:35:06 crc kubenswrapper[4692]: I0309 09:35:06.070103 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:35:06 crc kubenswrapper[4692]: I0309 09:35:06.928654 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.443750 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/openstack-galera-0"] Mar 09 09:35:09 crc kubenswrapper[4692]: E0309 09:35:09.443983 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="extract-utilities" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.443994 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="extract-utilities" Mar 09 09:35:09 crc kubenswrapper[4692]: E0309 09:35:09.444007 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerName="util" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444014 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerName="util" Mar 09 09:35:09 crc kubenswrapper[4692]: E0309 09:35:09.444022 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerName="extract" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444029 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerName="extract" Mar 09 09:35:09 crc kubenswrapper[4692]: E0309 09:35:09.444037 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="extract-content" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444043 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="extract-content" Mar 09 09:35:09 crc kubenswrapper[4692]: E0309 09:35:09.444053 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerName="pull" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444059 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerName="pull" Mar 09 09:35:09 crc kubenswrapper[4692]: E0309 09:35:09.444073 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="registry-server" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444079 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="registry-server" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444184 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df3c8b1-25ba-4b83-abdd-bf70af81208c" containerName="registry-server" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444198 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="381ff66a-6f6e-49c5-9c29-d414b7200dae" containerName="extract" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.444754 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.446811 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"openstack-scripts" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.447004 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"openshift-service-ca.crt" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.447074 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"kube-root-ca.crt" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.447200 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"galera-openstack-dockercfg-qsckl" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.448991 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"openstack-config-data" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.456896 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/openstack-galera-1"] Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.457966 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.462132 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-0"] Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.477308 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/openstack-galera-2"] Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.481591 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.484019 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-1"] Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.502998 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-2"] Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594429 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-config-data-default\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594490 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-config-data-default\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594514 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7b194e65-c0c6-4ab1-966d-99ce335b719b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594555 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqt8x\" (UniqueName: \"kubernetes.io/projected/7b194e65-c0c6-4ab1-966d-99ce335b719b-kube-api-access-hqt8x\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594583 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-operator-scripts\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594623 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594658 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-kolla-config\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594685 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.594956 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spgzr\" (UniqueName: \"kubernetes.io/projected/75f90081-7951-418d-9c49-5aa141ee85b0-kube-api-access-spgzr\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.595055 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-kolla-config\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.595181 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/75f90081-7951-418d-9c49-5aa141ee85b0-config-data-generated\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.595218 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.696735 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqt8x\" (UniqueName: \"kubernetes.io/projected/7b194e65-c0c6-4ab1-966d-99ce335b719b-kube-api-access-hqt8x\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.697075 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-operator-scripts\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.697111 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-kolla-config\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.697141 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5dq9\" (UniqueName: \"kubernetes.io/projected/87a076b1-43df-4290-903a-abc067ba6cad-kube-api-access-x5dq9\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.697180 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.697216 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.697241 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-kolla-config\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.698387 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-kolla-config\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.698529 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.698627 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spgzr\" (UniqueName: \"kubernetes.io/projected/75f90081-7951-418d-9c49-5aa141ee85b0-kube-api-access-spgzr\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.698715 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-kolla-config\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.698861 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-config-data-default\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.698972 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/87a076b1-43df-4290-903a-abc067ba6cad-config-data-generated\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699044 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") device mount path \"/mnt/openstack/pv01\"" pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699067 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.698981 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-operator-scripts\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699494 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-kolla-config\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699537 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/75f90081-7951-418d-9c49-5aa141ee85b0-config-data-generated\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699559 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699583 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-operator-scripts\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699609 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-config-data-default\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699675 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-config-data-default\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.699694 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7b194e65-c0c6-4ab1-966d-99ce335b719b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.700064 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7b194e65-c0c6-4ab1-966d-99ce335b719b-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.700332 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/75f90081-7951-418d-9c49-5aa141ee85b0-config-data-generated\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.700406 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") device mount path \"/mnt/openstack/pv04\"" pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.702038 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7b194e65-c0c6-4ab1-966d-99ce335b719b-config-data-default\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.702060 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/75f90081-7951-418d-9c49-5aa141ee85b0-config-data-default\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.729466 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqt8x\" (UniqueName: \"kubernetes.io/projected/7b194e65-c0c6-4ab1-966d-99ce335b719b-kube-api-access-hqt8x\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.746374 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"7b194e65-c0c6-4ab1-966d-99ce335b719b\") " pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.752861 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spgzr\" (UniqueName: \"kubernetes.io/projected/75f90081-7951-418d-9c49-5aa141ee85b0-kube-api-access-spgzr\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.763770 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.780057 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-1\" (UID: \"75f90081-7951-418d-9c49-5aa141ee85b0\") " pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.809441 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5dq9\" (UniqueName: \"kubernetes.io/projected/87a076b1-43df-4290-903a-abc067ba6cad-kube-api-access-x5dq9\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.809743 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.809905 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-config-data-default\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.809984 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/87a076b1-43df-4290-903a-abc067ba6cad-config-data-generated\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.810069 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-operator-scripts\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.810186 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-kolla-config\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.811011 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-kolla-config\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.816648 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") device mount path \"/mnt/openstack/pv06\"" pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.819610 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/87a076b1-43df-4290-903a-abc067ba6cad-config-data-generated\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.820360 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-config-data-default\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.824788 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/87a076b1-43df-4290-903a-abc067ba6cad-operator-scripts\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.869112 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5dq9\" (UniqueName: \"kubernetes.io/projected/87a076b1-43df-4290-903a-abc067ba6cad-kube-api-access-x5dq9\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.879274 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"87a076b1-43df-4290-903a-abc067ba6cad\") " pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.937260 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss"] Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.938451 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.941105 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.941317 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-5vhnv" Mar 09 09:35:09 crc kubenswrapper[4692]: I0309 09:35:09.957088 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss"] Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.085974 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.098560 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.117962 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac73a80c-7cf2-44eb-b384-dec4e2b20783-apiservice-cert\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.118084 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bzc9\" (UniqueName: \"kubernetes.io/projected/ac73a80c-7cf2-44eb-b384-dec4e2b20783-kube-api-access-5bzc9\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.118111 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac73a80c-7cf2-44eb-b384-dec4e2b20783-webhook-cert\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.161148 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-0"] Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.219309 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bzc9\" (UniqueName: \"kubernetes.io/projected/ac73a80c-7cf2-44eb-b384-dec4e2b20783-kube-api-access-5bzc9\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.219353 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac73a80c-7cf2-44eb-b384-dec4e2b20783-webhook-cert\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.219443 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac73a80c-7cf2-44eb-b384-dec4e2b20783-apiservice-cert\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.224795 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ac73a80c-7cf2-44eb-b384-dec4e2b20783-webhook-cert\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.224847 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ac73a80c-7cf2-44eb-b384-dec4e2b20783-apiservice-cert\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.237411 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bzc9\" (UniqueName: \"kubernetes.io/projected/ac73a80c-7cf2-44eb-b384-dec4e2b20783-kube-api-access-5bzc9\") pod \"infra-operator-controller-manager-5d6bc4dc48-7m6ss\" (UID: \"ac73a80c-7cf2-44eb-b384-dec4e2b20783\") " pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.297003 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.546036 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-1"] Mar 09 09:35:10 crc kubenswrapper[4692]: W0309 09:35:10.556263 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75f90081_7951_418d_9c49_5aa141ee85b0.slice/crio-7494c609e1c8eb62fef7818d513374769d25040627b6278d706a856ccf317ea6 WatchSource:0}: Error finding container 7494c609e1c8eb62fef7818d513374769d25040627b6278d706a856ccf317ea6: Status 404 returned error can't find the container with id 7494c609e1c8eb62fef7818d513374769d25040627b6278d706a856ccf317ea6 Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.634422 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/openstack-galera-2"] Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.789205 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss"] Mar 09 09:35:10 crc kubenswrapper[4692]: W0309 09:35:10.796181 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac73a80c_7cf2_44eb_b384_dec4e2b20783.slice/crio-1782bfe457bececf2691534323f0ee37b41851bbf2a281d65507e5b21b4e496a WatchSource:0}: Error finding container 1782bfe457bececf2691534323f0ee37b41851bbf2a281d65507e5b21b4e496a: Status 404 returned error can't find the container with id 1782bfe457bececf2691534323f0ee37b41851bbf2a281d65507e5b21b4e496a Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.916418 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"75f90081-7951-418d-9c49-5aa141ee85b0","Type":"ContainerStarted","Data":"7494c609e1c8eb62fef7818d513374769d25040627b6278d706a856ccf317ea6"} Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.917547 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"7b194e65-c0c6-4ab1-966d-99ce335b719b","Type":"ContainerStarted","Data":"39402fed81bc8a8ae9f2da0a941a3ca153afd8c36cb9665b2441266229df6fa8"} Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.918788 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"87a076b1-43df-4290-903a-abc067ba6cad","Type":"ContainerStarted","Data":"addce96ede469389590e670815fddf6567166c75f8380e8c82c6d146ca1a238a"} Mar 09 09:35:10 crc kubenswrapper[4692]: I0309 09:35:10.920634 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" event={"ID":"ac73a80c-7cf2-44eb-b384-dec4e2b20783","Type":"ContainerStarted","Data":"1782bfe457bececf2691534323f0ee37b41851bbf2a281d65507e5b21b4e496a"} Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.696328 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-884mk"] Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.697806 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.705385 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-884mk"] Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.853376 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57n99\" (UniqueName: \"kubernetes.io/projected/adec06e7-a23a-408e-8843-9a27379a00c1-kube-api-access-57n99\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.853528 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-utilities\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.853738 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-catalog-content\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.955712 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57n99\" (UniqueName: \"kubernetes.io/projected/adec06e7-a23a-408e-8843-9a27379a00c1-kube-api-access-57n99\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.955793 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-utilities\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.955834 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-catalog-content\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.956821 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-utilities\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.956942 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-catalog-content\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:11 crc kubenswrapper[4692]: I0309 09:35:11.983956 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57n99\" (UniqueName: \"kubernetes.io/projected/adec06e7-a23a-408e-8843-9a27379a00c1-kube-api-access-57n99\") pod \"certified-operators-884mk\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:12 crc kubenswrapper[4692]: I0309 09:35:12.025508 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:12 crc kubenswrapper[4692]: I0309 09:35:12.610651 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-884mk"] Mar 09 09:35:12 crc kubenswrapper[4692]: I0309 09:35:12.949354 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-884mk" event={"ID":"adec06e7-a23a-408e-8843-9a27379a00c1","Type":"ContainerStarted","Data":"7abdfa2bd56077ab5c0e7348838a2158045401e45e64835f8caf1a6fe08bdb66"} Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.076399 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zl49"] Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.076665 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6zl49" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="registry-server" containerID="cri-o://065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778" gracePeriod=2 Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.590892 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.694153 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tnnl\" (UniqueName: \"kubernetes.io/projected/f5a30e59-e0e0-4a43-be50-474183fb6db9-kube-api-access-5tnnl\") pod \"f5a30e59-e0e0-4a43-be50-474183fb6db9\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.694584 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-catalog-content\") pod \"f5a30e59-e0e0-4a43-be50-474183fb6db9\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.694621 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-utilities\") pod \"f5a30e59-e0e0-4a43-be50-474183fb6db9\" (UID: \"f5a30e59-e0e0-4a43-be50-474183fb6db9\") " Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.695669 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-utilities" (OuterVolumeSpecName: "utilities") pod "f5a30e59-e0e0-4a43-be50-474183fb6db9" (UID: "f5a30e59-e0e0-4a43-be50-474183fb6db9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.699696 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a30e59-e0e0-4a43-be50-474183fb6db9-kube-api-access-5tnnl" (OuterVolumeSpecName: "kube-api-access-5tnnl") pod "f5a30e59-e0e0-4a43-be50-474183fb6db9" (UID: "f5a30e59-e0e0-4a43-be50-474183fb6db9"). InnerVolumeSpecName "kube-api-access-5tnnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.737571 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5a30e59-e0e0-4a43-be50-474183fb6db9" (UID: "f5a30e59-e0e0-4a43-be50-474183fb6db9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.797267 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.797303 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5a30e59-e0e0-4a43-be50-474183fb6db9-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.797318 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tnnl\" (UniqueName: \"kubernetes.io/projected/f5a30e59-e0e0-4a43-be50-474183fb6db9-kube-api-access-5tnnl\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.962092 4692 generic.go:334] "Generic (PLEG): container finished" podID="adec06e7-a23a-408e-8843-9a27379a00c1" containerID="fb7929b55a68b2f3ad9b02578452211b01c6fcdf969d3d32c42474e217057704" exitCode=0 Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.962192 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-884mk" event={"ID":"adec06e7-a23a-408e-8843-9a27379a00c1","Type":"ContainerDied","Data":"fb7929b55a68b2f3ad9b02578452211b01c6fcdf969d3d32c42474e217057704"} Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.974777 4692 generic.go:334] "Generic (PLEG): container finished" podID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerID="065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778" exitCode=0 Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.974825 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zl49" event={"ID":"f5a30e59-e0e0-4a43-be50-474183fb6db9","Type":"ContainerDied","Data":"065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778"} Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.974856 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6zl49" event={"ID":"f5a30e59-e0e0-4a43-be50-474183fb6db9","Type":"ContainerDied","Data":"8eeb52651b16e7cb5bcc92bc38003bb5ce2fb358cac5378c2a8b50a9db27cc7e"} Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.974876 4692 scope.go:117] "RemoveContainer" containerID="065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778" Mar 09 09:35:13 crc kubenswrapper[4692]: I0309 09:35:13.975651 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6zl49" Mar 09 09:35:14 crc kubenswrapper[4692]: I0309 09:35:14.033192 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zl49"] Mar 09 09:35:14 crc kubenswrapper[4692]: I0309 09:35:14.033256 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6zl49"] Mar 09 09:35:14 crc kubenswrapper[4692]: I0309 09:35:14.081495 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" path="/var/lib/kubelet/pods/f5a30e59-e0e0-4a43-be50-474183fb6db9/volumes" Mar 09 09:35:18 crc kubenswrapper[4692]: I0309 09:35:18.319252 4692 scope.go:117] "RemoveContainer" containerID="63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893" Mar 09 09:35:23 crc kubenswrapper[4692]: I0309 09:35:23.151517 4692 scope.go:117] "RemoveContainer" containerID="dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62" Mar 09 09:35:23 crc kubenswrapper[4692]: I0309 09:35:23.197037 4692 scope.go:117] "RemoveContainer" containerID="065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778" Mar 09 09:35:23 crc kubenswrapper[4692]: E0309 09:35:23.197553 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778\": container with ID starting with 065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778 not found: ID does not exist" containerID="065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778" Mar 09 09:35:23 crc kubenswrapper[4692]: I0309 09:35:23.197594 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778"} err="failed to get container status \"065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778\": rpc error: code = NotFound desc = could not find container \"065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778\": container with ID starting with 065782fedfb3eb7f5f3fda07df3d8336198cde5f6231daf14d5a6fe3a1004778 not found: ID does not exist" Mar 09 09:35:23 crc kubenswrapper[4692]: I0309 09:35:23.197629 4692 scope.go:117] "RemoveContainer" containerID="63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893" Mar 09 09:35:23 crc kubenswrapper[4692]: E0309 09:35:23.197944 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893\": container with ID starting with 63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893 not found: ID does not exist" containerID="63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893" Mar 09 09:35:23 crc kubenswrapper[4692]: I0309 09:35:23.197981 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893"} err="failed to get container status \"63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893\": rpc error: code = NotFound desc = could not find container \"63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893\": container with ID starting with 63100159b155ec62f25c37d0855b3ca1823c185a0eea669089cf271f8f39d893 not found: ID does not exist" Mar 09 09:35:23 crc kubenswrapper[4692]: I0309 09:35:23.197999 4692 scope.go:117] "RemoveContainer" containerID="dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62" Mar 09 09:35:23 crc kubenswrapper[4692]: E0309 09:35:23.198242 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62\": container with ID starting with dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62 not found: ID does not exist" containerID="dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62" Mar 09 09:35:23 crc kubenswrapper[4692]: I0309 09:35:23.198268 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62"} err="failed to get container status \"dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62\": rpc error: code = NotFound desc = could not find container \"dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62\": container with ID starting with dd3ef0139e0a49bead27afed085c8ab7688c42e56aeab332cd8764d9432bff62 not found: ID does not exist" Mar 09 09:35:24 crc kubenswrapper[4692]: I0309 09:35:24.083419 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"87a076b1-43df-4290-903a-abc067ba6cad","Type":"ContainerStarted","Data":"9e8ee98bec000a189f08f0c5c66e3db05a75c4b5818d1ed00faf3ce6aceb1578"} Mar 09 09:35:24 crc kubenswrapper[4692]: I0309 09:35:24.083733 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" event={"ID":"ac73a80c-7cf2-44eb-b384-dec4e2b20783","Type":"ContainerStarted","Data":"dae9fee7a8b79797a53b9eb1556d6c7d8696aec354a922881929a44c372f8259"} Mar 09 09:35:24 crc kubenswrapper[4692]: I0309 09:35:24.083773 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:24 crc kubenswrapper[4692]: I0309 09:35:24.086455 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"75f90081-7951-418d-9c49-5aa141ee85b0","Type":"ContainerStarted","Data":"2c6ab34e6cb35fec46b42e6da00cce0056c6b1ae73738a0dcd52bf4797a91354"} Mar 09 09:35:24 crc kubenswrapper[4692]: I0309 09:35:24.088794 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"7b194e65-c0c6-4ab1-966d-99ce335b719b","Type":"ContainerStarted","Data":"72cc65e15a20c98c27e96b7a9e986e1a4c8841b79c26eaa4d415ec8c31b9a384"} Mar 09 09:35:24 crc kubenswrapper[4692]: I0309 09:35:24.145847 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" podStartSLOduration=2.797738825 podStartE2EDuration="15.145830885s" podCreationTimestamp="2026-03-09 09:35:09 +0000 UTC" firstStartedPulling="2026-03-09 09:35:10.804190121 +0000 UTC m=+911.628925702" lastFinishedPulling="2026-03-09 09:35:23.152282181 +0000 UTC m=+923.977017762" observedRunningTime="2026-03-09 09:35:24.141612175 +0000 UTC m=+924.966347746" watchObservedRunningTime="2026-03-09 09:35:24.145830885 +0000 UTC m=+924.970566456" Mar 09 09:35:25 crc kubenswrapper[4692]: I0309 09:35:25.097273 4692 generic.go:334] "Generic (PLEG): container finished" podID="adec06e7-a23a-408e-8843-9a27379a00c1" containerID="ab935e7cc9ab280d597980843cd40e532103a614cc31fa89fbfd5973409012ac" exitCode=0 Mar 09 09:35:25 crc kubenswrapper[4692]: I0309 09:35:25.097422 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-884mk" event={"ID":"adec06e7-a23a-408e-8843-9a27379a00c1","Type":"ContainerDied","Data":"ab935e7cc9ab280d597980843cd40e532103a614cc31fa89fbfd5973409012ac"} Mar 09 09:35:26 crc kubenswrapper[4692]: I0309 09:35:26.123203 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-884mk" event={"ID":"adec06e7-a23a-408e-8843-9a27379a00c1","Type":"ContainerStarted","Data":"e61cfaae6f5ff6a9ec37e064c9e6a71d267a913727e97099a7f6578d4de07a86"} Mar 09 09:35:26 crc kubenswrapper[4692]: I0309 09:35:26.141862 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-884mk" podStartSLOduration=4.5549001140000005 podStartE2EDuration="15.141845186s" podCreationTimestamp="2026-03-09 09:35:11 +0000 UTC" firstStartedPulling="2026-03-09 09:35:15.073914564 +0000 UTC m=+915.898650145" lastFinishedPulling="2026-03-09 09:35:25.660859626 +0000 UTC m=+926.485595217" observedRunningTime="2026-03-09 09:35:26.140488077 +0000 UTC m=+926.965223668" watchObservedRunningTime="2026-03-09 09:35:26.141845186 +0000 UTC m=+926.966580777" Mar 09 09:35:29 crc kubenswrapper[4692]: I0309 09:35:29.140896 4692 generic.go:334] "Generic (PLEG): container finished" podID="87a076b1-43df-4290-903a-abc067ba6cad" containerID="9e8ee98bec000a189f08f0c5c66e3db05a75c4b5818d1ed00faf3ce6aceb1578" exitCode=0 Mar 09 09:35:29 crc kubenswrapper[4692]: I0309 09:35:29.141009 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"87a076b1-43df-4290-903a-abc067ba6cad","Type":"ContainerDied","Data":"9e8ee98bec000a189f08f0c5c66e3db05a75c4b5818d1ed00faf3ce6aceb1578"} Mar 09 09:35:30 crc kubenswrapper[4692]: I0309 09:35:30.157543 4692 generic.go:334] "Generic (PLEG): container finished" podID="75f90081-7951-418d-9c49-5aa141ee85b0" containerID="2c6ab34e6cb35fec46b42e6da00cce0056c6b1ae73738a0dcd52bf4797a91354" exitCode=0 Mar 09 09:35:30 crc kubenswrapper[4692]: I0309 09:35:30.157631 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"75f90081-7951-418d-9c49-5aa141ee85b0","Type":"ContainerDied","Data":"2c6ab34e6cb35fec46b42e6da00cce0056c6b1ae73738a0dcd52bf4797a91354"} Mar 09 09:35:30 crc kubenswrapper[4692]: I0309 09:35:30.162733 4692 generic.go:334] "Generic (PLEG): container finished" podID="7b194e65-c0c6-4ab1-966d-99ce335b719b" containerID="72cc65e15a20c98c27e96b7a9e986e1a4c8841b79c26eaa4d415ec8c31b9a384" exitCode=0 Mar 09 09:35:30 crc kubenswrapper[4692]: I0309 09:35:30.162821 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"7b194e65-c0c6-4ab1-966d-99ce335b719b","Type":"ContainerDied","Data":"72cc65e15a20c98c27e96b7a9e986e1a4c8841b79c26eaa4d415ec8c31b9a384"} Mar 09 09:35:30 crc kubenswrapper[4692]: I0309 09:35:30.169091 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-2" event={"ID":"87a076b1-43df-4290-903a-abc067ba6cad","Type":"ContainerStarted","Data":"0fcd215902ef63ee69f97951a2191571cf65d20a990f8d078978f11136c571ce"} Mar 09 09:35:30 crc kubenswrapper[4692]: I0309 09:35:30.229294 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/openstack-galera-2" podStartSLOduration=9.398662786 podStartE2EDuration="22.229278231s" podCreationTimestamp="2026-03-09 09:35:08 +0000 UTC" firstStartedPulling="2026-03-09 09:35:10.642277872 +0000 UTC m=+911.467013453" lastFinishedPulling="2026-03-09 09:35:23.472893317 +0000 UTC m=+924.297628898" observedRunningTime="2026-03-09 09:35:30.225421511 +0000 UTC m=+931.050157102" watchObservedRunningTime="2026-03-09 09:35:30.229278231 +0000 UTC m=+931.054013812" Mar 09 09:35:30 crc kubenswrapper[4692]: I0309 09:35:30.303110 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5d6bc4dc48-7m6ss" Mar 09 09:35:31 crc kubenswrapper[4692]: I0309 09:35:31.176693 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-1" event={"ID":"75f90081-7951-418d-9c49-5aa141ee85b0","Type":"ContainerStarted","Data":"52878b72c03800572a0fc5c7f5e8772ed9701e5dd52ede518424b55996acbafd"} Mar 09 09:35:31 crc kubenswrapper[4692]: I0309 09:35:31.178655 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/openstack-galera-0" event={"ID":"7b194e65-c0c6-4ab1-966d-99ce335b719b","Type":"ContainerStarted","Data":"c0fa06c39b3c6b7145e4638b7b57c814370916d16993ed7d1f3ce08fc9971bee"} Mar 09 09:35:31 crc kubenswrapper[4692]: I0309 09:35:31.197943 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/openstack-galera-1" podStartSLOduration=10.237348151 podStartE2EDuration="23.197923974s" podCreationTimestamp="2026-03-09 09:35:08 +0000 UTC" firstStartedPulling="2026-03-09 09:35:10.558152492 +0000 UTC m=+911.382888083" lastFinishedPulling="2026-03-09 09:35:23.518728315 +0000 UTC m=+924.343463906" observedRunningTime="2026-03-09 09:35:31.194844516 +0000 UTC m=+932.019580087" watchObservedRunningTime="2026-03-09 09:35:31.197923974 +0000 UTC m=+932.022659555" Mar 09 09:35:31 crc kubenswrapper[4692]: I0309 09:35:31.223731 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/openstack-galera-0" podStartSLOduration=9.918853485 podStartE2EDuration="23.22371258s" podCreationTimestamp="2026-03-09 09:35:08 +0000 UTC" firstStartedPulling="2026-03-09 09:35:10.167328062 +0000 UTC m=+910.992063643" lastFinishedPulling="2026-03-09 09:35:23.472187157 +0000 UTC m=+924.296922738" observedRunningTime="2026-03-09 09:35:31.218980805 +0000 UTC m=+932.043716396" watchObservedRunningTime="2026-03-09 09:35:31.22371258 +0000 UTC m=+932.048448161" Mar 09 09:35:32 crc kubenswrapper[4692]: I0309 09:35:32.026182 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:32 crc kubenswrapper[4692]: I0309 09:35:32.026576 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:32 crc kubenswrapper[4692]: I0309 09:35:32.069840 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:32 crc kubenswrapper[4692]: I0309 09:35:32.234848 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.010335 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/memcached-0"] Mar 09 09:35:34 crc kubenswrapper[4692]: E0309 09:35:34.010822 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="extract-content" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.010833 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="extract-content" Mar 09 09:35:34 crc kubenswrapper[4692]: E0309 09:35:34.010847 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="extract-utilities" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.010852 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="extract-utilities" Mar 09 09:35:34 crc kubenswrapper[4692]: E0309 09:35:34.010869 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="registry-server" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.010877 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="registry-server" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.010978 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a30e59-e0e0-4a43-be50-474183fb6db9" containerName="registry-server" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.011426 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.013491 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"memcached-memcached-dockercfg-s7pkl" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.014354 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"memcached-config-data" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.025104 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/memcached-0"] Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.147492 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67894e62-3a7a-43a0-8497-74c261fbe1bb-config-data\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.147795 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/67894e62-3a7a-43a0-8497-74c261fbe1bb-kolla-config\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.147907 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnv68\" (UniqueName: \"kubernetes.io/projected/67894e62-3a7a-43a0-8497-74c261fbe1bb-kube-api-access-dnv68\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.249487 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67894e62-3a7a-43a0-8497-74c261fbe1bb-config-data\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.249744 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/67894e62-3a7a-43a0-8497-74c261fbe1bb-kolla-config\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.249869 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnv68\" (UniqueName: \"kubernetes.io/projected/67894e62-3a7a-43a0-8497-74c261fbe1bb-kube-api-access-dnv68\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.250556 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/67894e62-3a7a-43a0-8497-74c261fbe1bb-kolla-config\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.250568 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/67894e62-3a7a-43a0-8497-74c261fbe1bb-config-data\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.274472 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnv68\" (UniqueName: \"kubernetes.io/projected/67894e62-3a7a-43a0-8497-74c261fbe1bb-kube-api-access-dnv68\") pod \"memcached-0\" (UID: \"67894e62-3a7a-43a0-8497-74c261fbe1bb\") " pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.325238 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:34 crc kubenswrapper[4692]: I0309 09:35:34.731517 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/memcached-0"] Mar 09 09:35:34 crc kubenswrapper[4692]: W0309 09:35:34.735268 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67894e62_3a7a_43a0_8497_74c261fbe1bb.slice/crio-4630499e80634986902b4945ad00b2631e30a72d671cb6053e5b1dd0b8129764 WatchSource:0}: Error finding container 4630499e80634986902b4945ad00b2631e30a72d671cb6053e5b1dd0b8129764: Status 404 returned error can't find the container with id 4630499e80634986902b4945ad00b2631e30a72d671cb6053e5b1dd0b8129764 Mar 09 09:35:35 crc kubenswrapper[4692]: I0309 09:35:35.207848 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/memcached-0" event={"ID":"67894e62-3a7a-43a0-8497-74c261fbe1bb","Type":"ContainerStarted","Data":"4630499e80634986902b4945ad00b2631e30a72d671cb6053e5b1dd0b8129764"} Mar 09 09:35:36 crc kubenswrapper[4692]: I0309 09:35:36.080874 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-884mk"] Mar 09 09:35:36 crc kubenswrapper[4692]: I0309 09:35:36.082326 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-884mk" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="registry-server" containerID="cri-o://e61cfaae6f5ff6a9ec37e064c9e6a71d267a913727e97099a7f6578d4de07a86" gracePeriod=2 Mar 09 09:35:36 crc kubenswrapper[4692]: I0309 09:35:36.217875 4692 generic.go:334] "Generic (PLEG): container finished" podID="adec06e7-a23a-408e-8843-9a27379a00c1" containerID="e61cfaae6f5ff6a9ec37e064c9e6a71d267a913727e97099a7f6578d4de07a86" exitCode=0 Mar 09 09:35:36 crc kubenswrapper[4692]: I0309 09:35:36.217918 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-884mk" event={"ID":"adec06e7-a23a-408e-8843-9a27379a00c1","Type":"ContainerDied","Data":"e61cfaae6f5ff6a9ec37e064c9e6a71d267a913727e97099a7f6578d4de07a86"} Mar 09 09:35:36 crc kubenswrapper[4692]: I0309 09:35:36.936068 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.097967 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57n99\" (UniqueName: \"kubernetes.io/projected/adec06e7-a23a-408e-8843-9a27379a00c1-kube-api-access-57n99\") pod \"adec06e7-a23a-408e-8843-9a27379a00c1\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.098069 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-utilities\") pod \"adec06e7-a23a-408e-8843-9a27379a00c1\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.098095 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-catalog-content\") pod \"adec06e7-a23a-408e-8843-9a27379a00c1\" (UID: \"adec06e7-a23a-408e-8843-9a27379a00c1\") " Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.099341 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-utilities" (OuterVolumeSpecName: "utilities") pod "adec06e7-a23a-408e-8843-9a27379a00c1" (UID: "adec06e7-a23a-408e-8843-9a27379a00c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.105323 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adec06e7-a23a-408e-8843-9a27379a00c1-kube-api-access-57n99" (OuterVolumeSpecName: "kube-api-access-57n99") pod "adec06e7-a23a-408e-8843-9a27379a00c1" (UID: "adec06e7-a23a-408e-8843-9a27379a00c1"). InnerVolumeSpecName "kube-api-access-57n99". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.158627 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adec06e7-a23a-408e-8843-9a27379a00c1" (UID: "adec06e7-a23a-408e-8843-9a27379a00c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.199925 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57n99\" (UniqueName: \"kubernetes.io/projected/adec06e7-a23a-408e-8843-9a27379a00c1-kube-api-access-57n99\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.200173 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.200279 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adec06e7-a23a-408e-8843-9a27379a00c1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.227374 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-884mk" event={"ID":"adec06e7-a23a-408e-8843-9a27379a00c1","Type":"ContainerDied","Data":"7abdfa2bd56077ab5c0e7348838a2158045401e45e64835f8caf1a6fe08bdb66"} Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.227387 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-884mk" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.228240 4692 scope.go:117] "RemoveContainer" containerID="e61cfaae6f5ff6a9ec37e064c9e6a71d267a913727e97099a7f6578d4de07a86" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.229180 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/memcached-0" event={"ID":"67894e62-3a7a-43a0-8497-74c261fbe1bb","Type":"ContainerStarted","Data":"b47bc5a8b2de2316f4917ebe5df4c4b7c3ebe676ced9d4104910af656b369698"} Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.229333 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.253688 4692 scope.go:117] "RemoveContainer" containerID="ab935e7cc9ab280d597980843cd40e532103a614cc31fa89fbfd5973409012ac" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.261679 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/memcached-0" podStartSLOduration=2.340191202 podStartE2EDuration="4.261655936s" podCreationTimestamp="2026-03-09 09:35:33 +0000 UTC" firstStartedPulling="2026-03-09 09:35:34.736884361 +0000 UTC m=+935.561619942" lastFinishedPulling="2026-03-09 09:35:36.658349095 +0000 UTC m=+937.483084676" observedRunningTime="2026-03-09 09:35:37.247776661 +0000 UTC m=+938.072512242" watchObservedRunningTime="2026-03-09 09:35:37.261655936 +0000 UTC m=+938.086391517" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.276982 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-884mk"] Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.279300 4692 scope.go:117] "RemoveContainer" containerID="fb7929b55a68b2f3ad9b02578452211b01c6fcdf969d3d32c42474e217057704" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.285781 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-884mk"] Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.882541 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-4qdpt"] Mar 09 09:35:37 crc kubenswrapper[4692]: E0309 09:35:37.883099 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="extract-content" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.883110 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="extract-content" Mar 09 09:35:37 crc kubenswrapper[4692]: E0309 09:35:37.883121 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="registry-server" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.883127 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="registry-server" Mar 09 09:35:37 crc kubenswrapper[4692]: E0309 09:35:37.883140 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="extract-utilities" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.883146 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="extract-utilities" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.883264 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" containerName="registry-server" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.883678 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.885888 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-l5pfh" Mar 09 09:35:37 crc kubenswrapper[4692]: I0309 09:35:37.894868 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-4qdpt"] Mar 09 09:35:38 crc kubenswrapper[4692]: I0309 09:35:38.014344 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5w6c\" (UniqueName: \"kubernetes.io/projected/eb637dd6-49f9-46a6-b3d8-c5e2f41921dc-kube-api-access-j5w6c\") pod \"rabbitmq-cluster-operator-index-4qdpt\" (UID: \"eb637dd6-49f9-46a6-b3d8-c5e2f41921dc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:38 crc kubenswrapper[4692]: I0309 09:35:38.079649 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adec06e7-a23a-408e-8843-9a27379a00c1" path="/var/lib/kubelet/pods/adec06e7-a23a-408e-8843-9a27379a00c1/volumes" Mar 09 09:35:38 crc kubenswrapper[4692]: I0309 09:35:38.116152 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5w6c\" (UniqueName: \"kubernetes.io/projected/eb637dd6-49f9-46a6-b3d8-c5e2f41921dc-kube-api-access-j5w6c\") pod \"rabbitmq-cluster-operator-index-4qdpt\" (UID: \"eb637dd6-49f9-46a6-b3d8-c5e2f41921dc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:38 crc kubenswrapper[4692]: I0309 09:35:38.135225 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5w6c\" (UniqueName: \"kubernetes.io/projected/eb637dd6-49f9-46a6-b3d8-c5e2f41921dc-kube-api-access-j5w6c\") pod \"rabbitmq-cluster-operator-index-4qdpt\" (UID: \"eb637dd6-49f9-46a6-b3d8-c5e2f41921dc\") " pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:38 crc kubenswrapper[4692]: I0309 09:35:38.200079 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:38 crc kubenswrapper[4692]: I0309 09:35:38.612154 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-4qdpt"] Mar 09 09:35:38 crc kubenswrapper[4692]: W0309 09:35:38.622380 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb637dd6_49f9_46a6_b3d8_c5e2f41921dc.slice/crio-8ffd81843c766ff87c7811a14443936e44a1edeacd413328242a38366f90be95 WatchSource:0}: Error finding container 8ffd81843c766ff87c7811a14443936e44a1edeacd413328242a38366f90be95: Status 404 returned error can't find the container with id 8ffd81843c766ff87c7811a14443936e44a1edeacd413328242a38366f90be95 Mar 09 09:35:39 crc kubenswrapper[4692]: I0309 09:35:39.254002 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" event={"ID":"eb637dd6-49f9-46a6-b3d8-c5e2f41921dc","Type":"ContainerStarted","Data":"8ffd81843c766ff87c7811a14443936e44a1edeacd413328242a38366f90be95"} Mar 09 09:35:39 crc kubenswrapper[4692]: I0309 09:35:39.764946 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:39 crc kubenswrapper[4692]: I0309 09:35:39.765015 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:40 crc kubenswrapper[4692]: I0309 09:35:40.086638 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:40 crc kubenswrapper[4692]: I0309 09:35:40.086692 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:40 crc kubenswrapper[4692]: I0309 09:35:40.099574 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:40 crc kubenswrapper[4692]: I0309 09:35:40.099684 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:40 crc kubenswrapper[4692]: I0309 09:35:40.334928 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:41 crc kubenswrapper[4692]: I0309 09:35:41.418848 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/openstack-galera-2" Mar 09 09:35:44 crc kubenswrapper[4692]: I0309 09:35:44.291182 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" event={"ID":"eb637dd6-49f9-46a6-b3d8-c5e2f41921dc","Type":"ContainerStarted","Data":"c92d4b5c29f64d11bb4bf1288f5f0edf4c0e9884fdd8589810d43b1298bc25c2"} Mar 09 09:35:44 crc kubenswrapper[4692]: I0309 09:35:44.308287 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" podStartSLOduration=2.760597235 podStartE2EDuration="7.308269629s" podCreationTimestamp="2026-03-09 09:35:37 +0000 UTC" firstStartedPulling="2026-03-09 09:35:38.623875938 +0000 UTC m=+939.448611519" lastFinishedPulling="2026-03-09 09:35:43.171548332 +0000 UTC m=+943.996283913" observedRunningTime="2026-03-09 09:35:44.306028595 +0000 UTC m=+945.130764186" watchObservedRunningTime="2026-03-09 09:35:44.308269629 +0000 UTC m=+945.133005210" Mar 09 09:35:44 crc kubenswrapper[4692]: I0309 09:35:44.327335 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/memcached-0" Mar 09 09:35:47 crc kubenswrapper[4692]: I0309 09:35:47.624404 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:35:47 crc kubenswrapper[4692]: I0309 09:35:47.624991 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.201075 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.201129 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.228498 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.335464 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-4qdpt" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.548115 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/root-account-create-update-xp2sq"] Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.549294 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.555299 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"openstack-mariadb-root-db-secret" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.565731 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/root-account-create-update-xp2sq"] Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.656205 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45zr4\" (UniqueName: \"kubernetes.io/projected/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-kube-api-access-45zr4\") pod \"root-account-create-update-xp2sq\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.656272 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-operator-scripts\") pod \"root-account-create-update-xp2sq\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.757804 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45zr4\" (UniqueName: \"kubernetes.io/projected/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-kube-api-access-45zr4\") pod \"root-account-create-update-xp2sq\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.757886 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-operator-scripts\") pod \"root-account-create-update-xp2sq\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.758763 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-operator-scripts\") pod \"root-account-create-update-xp2sq\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.777534 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45zr4\" (UniqueName: \"kubernetes.io/projected/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-kube-api-access-45zr4\") pod \"root-account-create-update-xp2sq\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:48 crc kubenswrapper[4692]: I0309 09:35:48.869816 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:49 crc kubenswrapper[4692]: I0309 09:35:49.298099 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/root-account-create-update-xp2sq"] Mar 09 09:35:49 crc kubenswrapper[4692]: I0309 09:35:49.325973 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-xp2sq" event={"ID":"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1","Type":"ContainerStarted","Data":"bc9fedfb00a2e3be78def0d3e7e207a88c3daf2738bb0897e6262cce986d6053"} Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.182109 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="swift-kuttl-tests/openstack-galera-2" podUID="87a076b1-43df-4290-903a-abc067ba6cad" containerName="galera" probeResult="failure" output=< Mar 09 09:35:50 crc kubenswrapper[4692]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Mar 09 09:35:50 crc kubenswrapper[4692]: > Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.322555 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4"] Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.324636 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.327616 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lthqw" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.328415 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4"] Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.333559 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-xp2sq" event={"ID":"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1","Type":"ContainerStarted","Data":"138c618fb48aac30a1bb3124b9eb5bd1f25f2df24b7e57329768d7535fad9638"} Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.363243 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/root-account-create-update-xp2sq" podStartSLOduration=2.363224623 podStartE2EDuration="2.363224623s" podCreationTimestamp="2026-03-09 09:35:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:35:50.361131063 +0000 UTC m=+951.185866644" watchObservedRunningTime="2026-03-09 09:35:50.363224623 +0000 UTC m=+951.187960194" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.390294 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.390418 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d86f\" (UniqueName: \"kubernetes.io/projected/e6478415-5d9c-44fa-aed4-da919022fa54-kube-api-access-5d86f\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.390471 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.491128 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d86f\" (UniqueName: \"kubernetes.io/projected/e6478415-5d9c-44fa-aed4-da919022fa54-kube-api-access-5d86f\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.491196 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.491242 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.491729 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.492065 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.514498 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d86f\" (UniqueName: \"kubernetes.io/projected/e6478415-5d9c-44fa-aed4-da919022fa54-kube-api-access-5d86f\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:50 crc kubenswrapper[4692]: I0309 09:35:50.641316 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:51 crc kubenswrapper[4692]: I0309 09:35:51.103056 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4"] Mar 09 09:35:51 crc kubenswrapper[4692]: W0309 09:35:51.108001 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6478415_5d9c_44fa_aed4_da919022fa54.slice/crio-53e620ce9cb303a9770e4f74ef3ea4f0dd625012cc098977ef4c8d9610a4e076 WatchSource:0}: Error finding container 53e620ce9cb303a9770e4f74ef3ea4f0dd625012cc098977ef4c8d9610a4e076: Status 404 returned error can't find the container with id 53e620ce9cb303a9770e4f74ef3ea4f0dd625012cc098977ef4c8d9610a4e076 Mar 09 09:35:51 crc kubenswrapper[4692]: I0309 09:35:51.344249 4692 generic.go:334] "Generic (PLEG): container finished" podID="d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1" containerID="138c618fb48aac30a1bb3124b9eb5bd1f25f2df24b7e57329768d7535fad9638" exitCode=0 Mar 09 09:35:51 crc kubenswrapper[4692]: I0309 09:35:51.344295 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-xp2sq" event={"ID":"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1","Type":"ContainerDied","Data":"138c618fb48aac30a1bb3124b9eb5bd1f25f2df24b7e57329768d7535fad9638"} Mar 09 09:35:51 crc kubenswrapper[4692]: I0309 09:35:51.346237 4692 generic.go:334] "Generic (PLEG): container finished" podID="e6478415-5d9c-44fa-aed4-da919022fa54" containerID="453f0e824ea6631e36c11a2758e2aadb8ae8f045acbcf87bd1a5e7c171a3c5d8" exitCode=0 Mar 09 09:35:51 crc kubenswrapper[4692]: I0309 09:35:51.346281 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" event={"ID":"e6478415-5d9c-44fa-aed4-da919022fa54","Type":"ContainerDied","Data":"453f0e824ea6631e36c11a2758e2aadb8ae8f045acbcf87bd1a5e7c171a3c5d8"} Mar 09 09:35:51 crc kubenswrapper[4692]: I0309 09:35:51.346307 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" event={"ID":"e6478415-5d9c-44fa-aed4-da919022fa54","Type":"ContainerStarted","Data":"53e620ce9cb303a9770e4f74ef3ea4f0dd625012cc098977ef4c8d9610a4e076"} Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.354781 4692 generic.go:334] "Generic (PLEG): container finished" podID="e6478415-5d9c-44fa-aed4-da919022fa54" containerID="23754f197be866bca7721a1f52b2d2b5fb0f93b7619fe6674df16dc8f381e879" exitCode=0 Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.354826 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" event={"ID":"e6478415-5d9c-44fa-aed4-da919022fa54","Type":"ContainerDied","Data":"23754f197be866bca7721a1f52b2d2b5fb0f93b7619fe6674df16dc8f381e879"} Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.707136 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.823512 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45zr4\" (UniqueName: \"kubernetes.io/projected/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-kube-api-access-45zr4\") pod \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.823566 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-operator-scripts\") pod \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\" (UID: \"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1\") " Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.824517 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1" (UID: "d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.833341 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-kube-api-access-45zr4" (OuterVolumeSpecName: "kube-api-access-45zr4") pod "d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1" (UID: "d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1"). InnerVolumeSpecName "kube-api-access-45zr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.855883 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.925586 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45zr4\" (UniqueName: \"kubernetes.io/projected/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-kube-api-access-45zr4\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.926011 4692 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:52 crc kubenswrapper[4692]: I0309 09:35:52.941606 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/openstack-galera-0" Mar 09 09:35:53 crc kubenswrapper[4692]: I0309 09:35:53.363201 4692 generic.go:334] "Generic (PLEG): container finished" podID="e6478415-5d9c-44fa-aed4-da919022fa54" containerID="13d7b63fe98ee3b1bb247294a300b26b63166e573d4de9ffa6985d6edaeb9e83" exitCode=0 Mar 09 09:35:53 crc kubenswrapper[4692]: I0309 09:35:53.363236 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" event={"ID":"e6478415-5d9c-44fa-aed4-da919022fa54","Type":"ContainerDied","Data":"13d7b63fe98ee3b1bb247294a300b26b63166e573d4de9ffa6985d6edaeb9e83"} Mar 09 09:35:53 crc kubenswrapper[4692]: I0309 09:35:53.364743 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/root-account-create-update-xp2sq" event={"ID":"d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1","Type":"ContainerDied","Data":"bc9fedfb00a2e3be78def0d3e7e207a88c3daf2738bb0897e6262cce986d6053"} Mar 09 09:35:53 crc kubenswrapper[4692]: I0309 09:35:53.364801 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc9fedfb00a2e3be78def0d3e7e207a88c3daf2738bb0897e6262cce986d6053" Mar 09 09:35:53 crc kubenswrapper[4692]: I0309 09:35:53.364762 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/root-account-create-update-xp2sq" Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.653390 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.750545 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-util\") pod \"e6478415-5d9c-44fa-aed4-da919022fa54\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.750634 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-bundle\") pod \"e6478415-5d9c-44fa-aed4-da919022fa54\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.750670 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d86f\" (UniqueName: \"kubernetes.io/projected/e6478415-5d9c-44fa-aed4-da919022fa54-kube-api-access-5d86f\") pod \"e6478415-5d9c-44fa-aed4-da919022fa54\" (UID: \"e6478415-5d9c-44fa-aed4-da919022fa54\") " Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.752540 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-bundle" (OuterVolumeSpecName: "bundle") pod "e6478415-5d9c-44fa-aed4-da919022fa54" (UID: "e6478415-5d9c-44fa-aed4-da919022fa54"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.757335 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6478415-5d9c-44fa-aed4-da919022fa54-kube-api-access-5d86f" (OuterVolumeSpecName: "kube-api-access-5d86f") pod "e6478415-5d9c-44fa-aed4-da919022fa54" (UID: "e6478415-5d9c-44fa-aed4-da919022fa54"). InnerVolumeSpecName "kube-api-access-5d86f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.772092 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-util" (OuterVolumeSpecName: "util") pod "e6478415-5d9c-44fa-aed4-da919022fa54" (UID: "e6478415-5d9c-44fa-aed4-da919022fa54"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.852445 4692 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-util\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.852479 4692 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e6478415-5d9c-44fa-aed4-da919022fa54-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:54 crc kubenswrapper[4692]: I0309 09:35:54.852488 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d86f\" (UniqueName: \"kubernetes.io/projected/e6478415-5d9c-44fa-aed4-da919022fa54-kube-api-access-5d86f\") on node \"crc\" DevicePath \"\"" Mar 09 09:35:55 crc kubenswrapper[4692]: I0309 09:35:55.380288 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" event={"ID":"e6478415-5d9c-44fa-aed4-da919022fa54","Type":"ContainerDied","Data":"53e620ce9cb303a9770e4f74ef3ea4f0dd625012cc098977ef4c8d9610a4e076"} Mar 09 09:35:55 crc kubenswrapper[4692]: I0309 09:35:55.380647 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53e620ce9cb303a9770e4f74ef3ea4f0dd625012cc098977ef4c8d9610a4e076" Mar 09 09:35:55 crc kubenswrapper[4692]: I0309 09:35:55.380401 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4" Mar 09 09:35:55 crc kubenswrapper[4692]: I0309 09:35:55.820473 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:35:55 crc kubenswrapper[4692]: I0309 09:35:55.899047 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/openstack-galera-1" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.121973 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550816-j2x6f"] Mar 09 09:36:00 crc kubenswrapper[4692]: E0309 09:36:00.122470 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6478415-5d9c-44fa-aed4-da919022fa54" containerName="util" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.122481 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6478415-5d9c-44fa-aed4-da919022fa54" containerName="util" Mar 09 09:36:00 crc kubenswrapper[4692]: E0309 09:36:00.122493 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1" containerName="mariadb-account-create-update" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.122498 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1" containerName="mariadb-account-create-update" Mar 09 09:36:00 crc kubenswrapper[4692]: E0309 09:36:00.122512 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6478415-5d9c-44fa-aed4-da919022fa54" containerName="extract" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.122518 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6478415-5d9c-44fa-aed4-da919022fa54" containerName="extract" Mar 09 09:36:00 crc kubenswrapper[4692]: E0309 09:36:00.122528 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6478415-5d9c-44fa-aed4-da919022fa54" containerName="pull" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.122534 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6478415-5d9c-44fa-aed4-da919022fa54" containerName="pull" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.122635 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1" containerName="mariadb-account-create-update" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.122646 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6478415-5d9c-44fa-aed4-da919022fa54" containerName="extract" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.123014 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550816-j2x6f" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.126314 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.126487 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.126617 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.133939 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550816-j2x6f"] Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.224135 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsl68\" (UniqueName: \"kubernetes.io/projected/355a2d83-082d-4181-a67e-01373c53a52a-kube-api-access-wsl68\") pod \"auto-csr-approver-29550816-j2x6f\" (UID: \"355a2d83-082d-4181-a67e-01373c53a52a\") " pod="openshift-infra/auto-csr-approver-29550816-j2x6f" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.325220 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsl68\" (UniqueName: \"kubernetes.io/projected/355a2d83-082d-4181-a67e-01373c53a52a-kube-api-access-wsl68\") pod \"auto-csr-approver-29550816-j2x6f\" (UID: \"355a2d83-082d-4181-a67e-01373c53a52a\") " pod="openshift-infra/auto-csr-approver-29550816-j2x6f" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.343540 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsl68\" (UniqueName: \"kubernetes.io/projected/355a2d83-082d-4181-a67e-01373c53a52a-kube-api-access-wsl68\") pod \"auto-csr-approver-29550816-j2x6f\" (UID: \"355a2d83-082d-4181-a67e-01373c53a52a\") " pod="openshift-infra/auto-csr-approver-29550816-j2x6f" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.454211 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550816-j2x6f" Mar 09 09:36:00 crc kubenswrapper[4692]: I0309 09:36:00.723929 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550816-j2x6f"] Mar 09 09:36:01 crc kubenswrapper[4692]: I0309 09:36:01.414475 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550816-j2x6f" event={"ID":"355a2d83-082d-4181-a67e-01373c53a52a","Type":"ContainerStarted","Data":"0ad0f18122e171d28aae2b72b035bc0952a8145bd898800863532f5510d44089"} Mar 09 09:36:02 crc kubenswrapper[4692]: I0309 09:36:02.422685 4692 generic.go:334] "Generic (PLEG): container finished" podID="355a2d83-082d-4181-a67e-01373c53a52a" containerID="2428897e623dc39d6402064f77ddc16fe944452aeac8b1afdc908e25ff32a7e6" exitCode=0 Mar 09 09:36:02 crc kubenswrapper[4692]: I0309 09:36:02.422737 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550816-j2x6f" event={"ID":"355a2d83-082d-4181-a67e-01373c53a52a","Type":"ContainerDied","Data":"2428897e623dc39d6402064f77ddc16fe944452aeac8b1afdc908e25ff32a7e6"} Mar 09 09:36:03 crc kubenswrapper[4692]: I0309 09:36:03.805913 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550816-j2x6f" Mar 09 09:36:03 crc kubenswrapper[4692]: I0309 09:36:03.881279 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsl68\" (UniqueName: \"kubernetes.io/projected/355a2d83-082d-4181-a67e-01373c53a52a-kube-api-access-wsl68\") pod \"355a2d83-082d-4181-a67e-01373c53a52a\" (UID: \"355a2d83-082d-4181-a67e-01373c53a52a\") " Mar 09 09:36:03 crc kubenswrapper[4692]: I0309 09:36:03.900869 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/355a2d83-082d-4181-a67e-01373c53a52a-kube-api-access-wsl68" (OuterVolumeSpecName: "kube-api-access-wsl68") pod "355a2d83-082d-4181-a67e-01373c53a52a" (UID: "355a2d83-082d-4181-a67e-01373c53a52a"). InnerVolumeSpecName "kube-api-access-wsl68". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:36:03 crc kubenswrapper[4692]: I0309 09:36:03.983319 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsl68\" (UniqueName: \"kubernetes.io/projected/355a2d83-082d-4181-a67e-01373c53a52a-kube-api-access-wsl68\") on node \"crc\" DevicePath \"\"" Mar 09 09:36:04 crc kubenswrapper[4692]: I0309 09:36:04.436729 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550816-j2x6f" event={"ID":"355a2d83-082d-4181-a67e-01373c53a52a","Type":"ContainerDied","Data":"0ad0f18122e171d28aae2b72b035bc0952a8145bd898800863532f5510d44089"} Mar 09 09:36:04 crc kubenswrapper[4692]: I0309 09:36:04.436767 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ad0f18122e171d28aae2b72b035bc0952a8145bd898800863532f5510d44089" Mar 09 09:36:04 crc kubenswrapper[4692]: I0309 09:36:04.436802 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550816-j2x6f" Mar 09 09:36:04 crc kubenswrapper[4692]: I0309 09:36:04.858445 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550810-4ckmx"] Mar 09 09:36:04 crc kubenswrapper[4692]: I0309 09:36:04.862286 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550810-4ckmx"] Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.078520 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4acdeb1a-b335-45c9-9ba6-22fe29675aaf" path="/var/lib/kubelet/pods/4acdeb1a-b335-45c9-9ba6-22fe29675aaf/volumes" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.174879 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2"] Mar 09 09:36:06 crc kubenswrapper[4692]: E0309 09:36:06.175858 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="355a2d83-082d-4181-a67e-01373c53a52a" containerName="oc" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.175888 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="355a2d83-082d-4181-a67e-01373c53a52a" containerName="oc" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.176078 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="355a2d83-082d-4181-a67e-01373c53a52a" containerName="oc" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.176827 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.180143 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-xlmj9" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.190191 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2"] Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.211859 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crhdm\" (UniqueName: \"kubernetes.io/projected/335311a9-fa11-4fd5-b50e-29605ec7037e-kube-api-access-crhdm\") pod \"rabbitmq-cluster-operator-779fc9694b-7pps2\" (UID: \"335311a9-fa11-4fd5-b50e-29605ec7037e\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.313829 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crhdm\" (UniqueName: \"kubernetes.io/projected/335311a9-fa11-4fd5-b50e-29605ec7037e-kube-api-access-crhdm\") pod \"rabbitmq-cluster-operator-779fc9694b-7pps2\" (UID: \"335311a9-fa11-4fd5-b50e-29605ec7037e\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.344129 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crhdm\" (UniqueName: \"kubernetes.io/projected/335311a9-fa11-4fd5-b50e-29605ec7037e-kube-api-access-crhdm\") pod \"rabbitmq-cluster-operator-779fc9694b-7pps2\" (UID: \"335311a9-fa11-4fd5-b50e-29605ec7037e\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.492216 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" Mar 09 09:36:06 crc kubenswrapper[4692]: I0309 09:36:06.884274 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2"] Mar 09 09:36:07 crc kubenswrapper[4692]: I0309 09:36:07.459203 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" event={"ID":"335311a9-fa11-4fd5-b50e-29605ec7037e","Type":"ContainerStarted","Data":"5dbebb01fd2322c952b9b4f6ebac73b5d92aa7ba1789f966d1fcf19c787a4da2"} Mar 09 09:36:10 crc kubenswrapper[4692]: I0309 09:36:10.481323 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" event={"ID":"335311a9-fa11-4fd5-b50e-29605ec7037e","Type":"ContainerStarted","Data":"cc77c995546efdb026263b6dbf62ea96027a0de863ddeb5e3c77d615ba7d1114"} Mar 09 09:36:10 crc kubenswrapper[4692]: I0309 09:36:10.500674 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-7pps2" podStartSLOduration=1.310377214 podStartE2EDuration="4.500652555s" podCreationTimestamp="2026-03-09 09:36:06 +0000 UTC" firstStartedPulling="2026-03-09 09:36:06.895135398 +0000 UTC m=+967.719870979" lastFinishedPulling="2026-03-09 09:36:10.085410739 +0000 UTC m=+970.910146320" observedRunningTime="2026-03-09 09:36:10.495689883 +0000 UTC m=+971.320425464" watchObservedRunningTime="2026-03-09 09:36:10.500652555 +0000 UTC m=+971.325388146" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.706195 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/rabbitmq-server-0"] Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.707759 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.712257 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"rabbitmq-server-conf" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.712726 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"rabbitmq-plugins-conf" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.712807 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"rabbitmq-default-user" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.713210 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"rabbitmq-server-dockercfg-75vdt" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.713318 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"rabbitmq-erlang-cookie" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.727363 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/rabbitmq-server-0"] Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763138 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqww5\" (UniqueName: \"kubernetes.io/projected/8075e838-fcea-4974-aead-79a1c9f94aa9-kube-api-access-vqww5\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763204 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763260 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8075e838-fcea-4974-aead-79a1c9f94aa9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763306 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763329 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763367 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8075e838-fcea-4974-aead-79a1c9f94aa9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763388 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.763407 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8075e838-fcea-4974-aead-79a1c9f94aa9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.864932 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8075e838-fcea-4974-aead-79a1c9f94aa9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.864979 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865004 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8075e838-fcea-4974-aead-79a1c9f94aa9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865035 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqww5\" (UniqueName: \"kubernetes.io/projected/8075e838-fcea-4974-aead-79a1c9f94aa9-kube-api-access-vqww5\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865058 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865092 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8075e838-fcea-4974-aead-79a1c9f94aa9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865120 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865136 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865599 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.865686 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.866201 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8075e838-fcea-4974-aead-79a1c9f94aa9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.869086 4692 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.869351 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f7c4a185d407095719665fcfd11ed683f407ff255874108b362809d9a31fcd5a/globalmount\"" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.871666 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8075e838-fcea-4974-aead-79a1c9f94aa9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.871938 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8075e838-fcea-4974-aead-79a1c9f94aa9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.876678 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8075e838-fcea-4974-aead-79a1c9f94aa9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.884903 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqww5\" (UniqueName: \"kubernetes.io/projected/8075e838-fcea-4974-aead-79a1c9f94aa9-kube-api-access-vqww5\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:16 crc kubenswrapper[4692]: I0309 09:36:16.896585 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4067dd33-47f7-4beb-9b87-9eb2c5c87aca\") pod \"rabbitmq-server-0\" (UID: \"8075e838-fcea-4974-aead-79a1c9f94aa9\") " pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.027153 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.476145 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/rabbitmq-server-0"] Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.523215 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"8075e838-fcea-4974-aead-79a1c9f94aa9","Type":"ContainerStarted","Data":"b96902a6a00d2a84303aca46e3fd50abb6f14c71c7bbb3362a13bf9a563e99d3"} Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.624430 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.624493 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.684052 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-62dn4"] Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.684988 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-62dn4" Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.686889 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-sj4j8" Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.689987 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-62dn4"] Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.775331 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fcpr\" (UniqueName: \"kubernetes.io/projected/9c490aa7-e100-44b7-9379-0624c2579ae0-kube-api-access-9fcpr\") pod \"keystone-operator-index-62dn4\" (UID: \"9c490aa7-e100-44b7-9379-0624c2579ae0\") " pod="openstack-operators/keystone-operator-index-62dn4" Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.877303 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fcpr\" (UniqueName: \"kubernetes.io/projected/9c490aa7-e100-44b7-9379-0624c2579ae0-kube-api-access-9fcpr\") pod \"keystone-operator-index-62dn4\" (UID: \"9c490aa7-e100-44b7-9379-0624c2579ae0\") " pod="openstack-operators/keystone-operator-index-62dn4" Mar 09 09:36:17 crc kubenswrapper[4692]: I0309 09:36:17.909419 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fcpr\" (UniqueName: \"kubernetes.io/projected/9c490aa7-e100-44b7-9379-0624c2579ae0-kube-api-access-9fcpr\") pod \"keystone-operator-index-62dn4\" (UID: \"9c490aa7-e100-44b7-9379-0624c2579ae0\") " pod="openstack-operators/keystone-operator-index-62dn4" Mar 09 09:36:18 crc kubenswrapper[4692]: I0309 09:36:18.003574 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-62dn4" Mar 09 09:36:18 crc kubenswrapper[4692]: I0309 09:36:18.386376 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-62dn4"] Mar 09 09:36:18 crc kubenswrapper[4692]: W0309 09:36:18.393382 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c490aa7_e100_44b7_9379_0624c2579ae0.slice/crio-15311a5dd5fa365c77d6a34ea14be9ceb66459494d3431c02da690dbe132f223 WatchSource:0}: Error finding container 15311a5dd5fa365c77d6a34ea14be9ceb66459494d3431c02da690dbe132f223: Status 404 returned error can't find the container with id 15311a5dd5fa365c77d6a34ea14be9ceb66459494d3431c02da690dbe132f223 Mar 09 09:36:18 crc kubenswrapper[4692]: I0309 09:36:18.532582 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-62dn4" event={"ID":"9c490aa7-e100-44b7-9379-0624c2579ae0","Type":"ContainerStarted","Data":"15311a5dd5fa365c77d6a34ea14be9ceb66459494d3431c02da690dbe132f223"} Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.083239 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-62dn4"] Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.560413 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-62dn4" event={"ID":"9c490aa7-e100-44b7-9379-0624c2579ae0","Type":"ContainerStarted","Data":"5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f"} Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.560527 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-62dn4" podUID="9c490aa7-e100-44b7-9379-0624c2579ae0" containerName="registry-server" containerID="cri-o://5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f" gracePeriod=2 Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.579329 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-62dn4" podStartSLOduration=2.310561881 podStartE2EDuration="5.57930628s" podCreationTimestamp="2026-03-09 09:36:17 +0000 UTC" firstStartedPulling="2026-03-09 09:36:18.39666966 +0000 UTC m=+979.221405241" lastFinishedPulling="2026-03-09 09:36:21.665414059 +0000 UTC m=+982.490149640" observedRunningTime="2026-03-09 09:36:22.572482265 +0000 UTC m=+983.397217856" watchObservedRunningTime="2026-03-09 09:36:22.57930628 +0000 UTC m=+983.404041861" Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.690846 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-682hs"] Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.691944 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.698750 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-682hs"] Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.849379 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czc4p\" (UniqueName: \"kubernetes.io/projected/2f8455e8-81a6-427c-ad1b-9b63aed74e4d-kube-api-access-czc4p\") pod \"keystone-operator-index-682hs\" (UID: \"2f8455e8-81a6-427c-ad1b-9b63aed74e4d\") " pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.950401 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czc4p\" (UniqueName: \"kubernetes.io/projected/2f8455e8-81a6-427c-ad1b-9b63aed74e4d-kube-api-access-czc4p\") pod \"keystone-operator-index-682hs\" (UID: \"2f8455e8-81a6-427c-ad1b-9b63aed74e4d\") " pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:22 crc kubenswrapper[4692]: I0309 09:36:22.969992 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czc4p\" (UniqueName: \"kubernetes.io/projected/2f8455e8-81a6-427c-ad1b-9b63aed74e4d-kube-api-access-czc4p\") pod \"keystone-operator-index-682hs\" (UID: \"2f8455e8-81a6-427c-ad1b-9b63aed74e4d\") " pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.013487 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.157865 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-62dn4_9c490aa7-e100-44b7-9379-0624c2579ae0/registry-server/0.log" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.157951 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-62dn4" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.360043 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fcpr\" (UniqueName: \"kubernetes.io/projected/9c490aa7-e100-44b7-9379-0624c2579ae0-kube-api-access-9fcpr\") pod \"9c490aa7-e100-44b7-9379-0624c2579ae0\" (UID: \"9c490aa7-e100-44b7-9379-0624c2579ae0\") " Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.364584 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c490aa7-e100-44b7-9379-0624c2579ae0-kube-api-access-9fcpr" (OuterVolumeSpecName: "kube-api-access-9fcpr") pod "9c490aa7-e100-44b7-9379-0624c2579ae0" (UID: "9c490aa7-e100-44b7-9379-0624c2579ae0"). InnerVolumeSpecName "kube-api-access-9fcpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.461269 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fcpr\" (UniqueName: \"kubernetes.io/projected/9c490aa7-e100-44b7-9379-0624c2579ae0-kube-api-access-9fcpr\") on node \"crc\" DevicePath \"\"" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.526286 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-682hs"] Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.567963 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"8075e838-fcea-4974-aead-79a1c9f94aa9","Type":"ContainerStarted","Data":"41e2107404f52cb893a332d456b57063181f111eb2bec62a9c373fa5055b5bdc"} Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.569859 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-62dn4_9c490aa7-e100-44b7-9379-0624c2579ae0/registry-server/0.log" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.569896 4692 generic.go:334] "Generic (PLEG): container finished" podID="9c490aa7-e100-44b7-9379-0624c2579ae0" containerID="5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f" exitCode=1 Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.569969 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-62dn4" event={"ID":"9c490aa7-e100-44b7-9379-0624c2579ae0","Type":"ContainerDied","Data":"5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f"} Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.569986 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-62dn4" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.570042 4692 scope.go:117] "RemoveContainer" containerID="5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.570027 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-62dn4" event={"ID":"9c490aa7-e100-44b7-9379-0624c2579ae0","Type":"ContainerDied","Data":"15311a5dd5fa365c77d6a34ea14be9ceb66459494d3431c02da690dbe132f223"} Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.571278 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-682hs" event={"ID":"2f8455e8-81a6-427c-ad1b-9b63aed74e4d","Type":"ContainerStarted","Data":"db2d8e260b99e55cd99cc83a67d79ab58ae49f41bd5bc2c6ffc426433a16e29b"} Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.591675 4692 scope.go:117] "RemoveContainer" containerID="5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f" Mar 09 09:36:23 crc kubenswrapper[4692]: E0309 09:36:23.593524 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f\": container with ID starting with 5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f not found: ID does not exist" containerID="5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.593583 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f"} err="failed to get container status \"5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f\": rpc error: code = NotFound desc = could not find container \"5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f\": container with ID starting with 5ca33bc2e9d3f56ded908e28a2279880b8c8bacc324f1f448e6a8e7f1da49c2f not found: ID does not exist" Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.610382 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-62dn4"] Mar 09 09:36:23 crc kubenswrapper[4692]: I0309 09:36:23.616423 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-62dn4"] Mar 09 09:36:24 crc kubenswrapper[4692]: I0309 09:36:24.079048 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c490aa7-e100-44b7-9379-0624c2579ae0" path="/var/lib/kubelet/pods/9c490aa7-e100-44b7-9379-0624c2579ae0/volumes" Mar 09 09:36:24 crc kubenswrapper[4692]: I0309 09:36:24.579236 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-682hs" event={"ID":"2f8455e8-81a6-427c-ad1b-9b63aed74e4d","Type":"ContainerStarted","Data":"4a97abf33cacfd3677100a02a7e50ab41fff0680e516ac1adff416a2676d3ce5"} Mar 09 09:36:24 crc kubenswrapper[4692]: I0309 09:36:24.599605 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-682hs" podStartSLOduration=2.192913481 podStartE2EDuration="2.599582383s" podCreationTimestamp="2026-03-09 09:36:22 +0000 UTC" firstStartedPulling="2026-03-09 09:36:23.53491261 +0000 UTC m=+984.359648191" lastFinishedPulling="2026-03-09 09:36:23.941581512 +0000 UTC m=+984.766317093" observedRunningTime="2026-03-09 09:36:24.594256731 +0000 UTC m=+985.418992332" watchObservedRunningTime="2026-03-09 09:36:24.599582383 +0000 UTC m=+985.424317974" Mar 09 09:36:33 crc kubenswrapper[4692]: I0309 09:36:33.014113 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:33 crc kubenswrapper[4692]: I0309 09:36:33.014689 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:33 crc kubenswrapper[4692]: I0309 09:36:33.040214 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:33 crc kubenswrapper[4692]: I0309 09:36:33.677890 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-682hs" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.130738 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq"] Mar 09 09:36:36 crc kubenswrapper[4692]: E0309 09:36:36.131368 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c490aa7-e100-44b7-9379-0624c2579ae0" containerName="registry-server" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.131384 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c490aa7-e100-44b7-9379-0624c2579ae0" containerName="registry-server" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.131552 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c490aa7-e100-44b7-9379-0624c2579ae0" containerName="registry-server" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.133030 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.137268 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lthqw" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.139576 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq"] Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.265734 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95tfx\" (UniqueName: \"kubernetes.io/projected/093b8260-b333-4f8e-a8b2-8313868609be-kube-api-access-95tfx\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.265846 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.265900 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.366802 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95tfx\" (UniqueName: \"kubernetes.io/projected/093b8260-b333-4f8e-a8b2-8313868609be-kube-api-access-95tfx\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.366888 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.366926 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.367442 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.367632 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.403537 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95tfx\" (UniqueName: \"kubernetes.io/projected/093b8260-b333-4f8e-a8b2-8313868609be-kube-api-access-95tfx\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.456217 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:36 crc kubenswrapper[4692]: I0309 09:36:36.720473 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq"] Mar 09 09:36:36 crc kubenswrapper[4692]: W0309 09:36:36.726205 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod093b8260_b333_4f8e_a8b2_8313868609be.slice/crio-2989f009cfcf04d1a3c7703093dc89404e2561614a0a8b923fc1bd1a5d5ed341 WatchSource:0}: Error finding container 2989f009cfcf04d1a3c7703093dc89404e2561614a0a8b923fc1bd1a5d5ed341: Status 404 returned error can't find the container with id 2989f009cfcf04d1a3c7703093dc89404e2561614a0a8b923fc1bd1a5d5ed341 Mar 09 09:36:37 crc kubenswrapper[4692]: I0309 09:36:37.678341 4692 generic.go:334] "Generic (PLEG): container finished" podID="093b8260-b333-4f8e-a8b2-8313868609be" containerID="6c5c1bfa30988a773cc9546acd488699f50682f04395830f81711bf2d49366df" exitCode=0 Mar 09 09:36:37 crc kubenswrapper[4692]: I0309 09:36:37.678405 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" event={"ID":"093b8260-b333-4f8e-a8b2-8313868609be","Type":"ContainerDied","Data":"6c5c1bfa30988a773cc9546acd488699f50682f04395830f81711bf2d49366df"} Mar 09 09:36:37 crc kubenswrapper[4692]: I0309 09:36:37.678630 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" event={"ID":"093b8260-b333-4f8e-a8b2-8313868609be","Type":"ContainerStarted","Data":"2989f009cfcf04d1a3c7703093dc89404e2561614a0a8b923fc1bd1a5d5ed341"} Mar 09 09:36:38 crc kubenswrapper[4692]: I0309 09:36:38.686545 4692 generic.go:334] "Generic (PLEG): container finished" podID="093b8260-b333-4f8e-a8b2-8313868609be" containerID="064d1022bd93a4622c7eb9261a04651fe561919fed0247158dd6abd6cc9c9903" exitCode=0 Mar 09 09:36:38 crc kubenswrapper[4692]: I0309 09:36:38.686593 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" event={"ID":"093b8260-b333-4f8e-a8b2-8313868609be","Type":"ContainerDied","Data":"064d1022bd93a4622c7eb9261a04651fe561919fed0247158dd6abd6cc9c9903"} Mar 09 09:36:39 crc kubenswrapper[4692]: I0309 09:36:39.697658 4692 generic.go:334] "Generic (PLEG): container finished" podID="093b8260-b333-4f8e-a8b2-8313868609be" containerID="e87d89046b0405c7dddec6b245984d078a2622ed27079396bd9dd89fc4e646ec" exitCode=0 Mar 09 09:36:39 crc kubenswrapper[4692]: I0309 09:36:39.697701 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" event={"ID":"093b8260-b333-4f8e-a8b2-8313868609be","Type":"ContainerDied","Data":"e87d89046b0405c7dddec6b245984d078a2622ed27079396bd9dd89fc4e646ec"} Mar 09 09:36:40 crc kubenswrapper[4692]: I0309 09:36:40.952684 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.126530 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-bundle\") pod \"093b8260-b333-4f8e-a8b2-8313868609be\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.126632 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95tfx\" (UniqueName: \"kubernetes.io/projected/093b8260-b333-4f8e-a8b2-8313868609be-kube-api-access-95tfx\") pod \"093b8260-b333-4f8e-a8b2-8313868609be\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.126684 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-util\") pod \"093b8260-b333-4f8e-a8b2-8313868609be\" (UID: \"093b8260-b333-4f8e-a8b2-8313868609be\") " Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.127615 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-bundle" (OuterVolumeSpecName: "bundle") pod "093b8260-b333-4f8e-a8b2-8313868609be" (UID: "093b8260-b333-4f8e-a8b2-8313868609be"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.133051 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093b8260-b333-4f8e-a8b2-8313868609be-kube-api-access-95tfx" (OuterVolumeSpecName: "kube-api-access-95tfx") pod "093b8260-b333-4f8e-a8b2-8313868609be" (UID: "093b8260-b333-4f8e-a8b2-8313868609be"). InnerVolumeSpecName "kube-api-access-95tfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.143881 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-util" (OuterVolumeSpecName: "util") pod "093b8260-b333-4f8e-a8b2-8313868609be" (UID: "093b8260-b333-4f8e-a8b2-8313868609be"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.227669 4692 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.227699 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95tfx\" (UniqueName: \"kubernetes.io/projected/093b8260-b333-4f8e-a8b2-8313868609be-kube-api-access-95tfx\") on node \"crc\" DevicePath \"\"" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.227711 4692 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/093b8260-b333-4f8e-a8b2-8313868609be-util\") on node \"crc\" DevicePath \"\"" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.709326 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" event={"ID":"093b8260-b333-4f8e-a8b2-8313868609be","Type":"ContainerDied","Data":"2989f009cfcf04d1a3c7703093dc89404e2561614a0a8b923fc1bd1a5d5ed341"} Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.709695 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2989f009cfcf04d1a3c7703093dc89404e2561614a0a8b923fc1bd1a5d5ed341" Mar 09 09:36:41 crc kubenswrapper[4692]: I0309 09:36:41.709772 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq" Mar 09 09:36:47 crc kubenswrapper[4692]: I0309 09:36:47.623589 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:36:47 crc kubenswrapper[4692]: I0309 09:36:47.624232 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:36:47 crc kubenswrapper[4692]: I0309 09:36:47.624293 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:36:47 crc kubenswrapper[4692]: I0309 09:36:47.624944 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7bb4acc73ba36420c0329ef9cb241fb590012d44ec365a56134ad986b42e9b2"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:36:47 crc kubenswrapper[4692]: I0309 09:36:47.625010 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://c7bb4acc73ba36420c0329ef9cb241fb590012d44ec365a56134ad986b42e9b2" gracePeriod=600 Mar 09 09:36:48 crc kubenswrapper[4692]: I0309 09:36:48.750135 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="c7bb4acc73ba36420c0329ef9cb241fb590012d44ec365a56134ad986b42e9b2" exitCode=0 Mar 09 09:36:48 crc kubenswrapper[4692]: I0309 09:36:48.750204 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"c7bb4acc73ba36420c0329ef9cb241fb590012d44ec365a56134ad986b42e9b2"} Mar 09 09:36:48 crc kubenswrapper[4692]: I0309 09:36:48.750517 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"a00e663af9bd3bc83fcf84afc788c06c37b9025a209fe4f4ecbe66f9658cebf6"} Mar 09 09:36:48 crc kubenswrapper[4692]: I0309 09:36:48.750535 4692 scope.go:117] "RemoveContainer" containerID="5fa3fd1dcbe9393b3fc6c43e0491320471ea475c9f05b37486ae66af62282e34" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.069154 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb"] Mar 09 09:36:51 crc kubenswrapper[4692]: E0309 09:36:51.070051 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093b8260-b333-4f8e-a8b2-8313868609be" containerName="extract" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.070064 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="093b8260-b333-4f8e-a8b2-8313868609be" containerName="extract" Mar 09 09:36:51 crc kubenswrapper[4692]: E0309 09:36:51.070076 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093b8260-b333-4f8e-a8b2-8313868609be" containerName="pull" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.070081 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="093b8260-b333-4f8e-a8b2-8313868609be" containerName="pull" Mar 09 09:36:51 crc kubenswrapper[4692]: E0309 09:36:51.070091 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093b8260-b333-4f8e-a8b2-8313868609be" containerName="util" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.070097 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="093b8260-b333-4f8e-a8b2-8313868609be" containerName="util" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.070238 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="093b8260-b333-4f8e-a8b2-8313868609be" containerName="extract" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.070702 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.072060 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.073955 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-ns9jk" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.080421 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d405d006-df42-4994-a407-0c3779dcf45c-apiservice-cert\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.080489 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d405d006-df42-4994-a407-0c3779dcf45c-webhook-cert\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.080533 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fj85\" (UniqueName: \"kubernetes.io/projected/d405d006-df42-4994-a407-0c3779dcf45c-kube-api-access-6fj85\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.083570 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb"] Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.182082 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d405d006-df42-4994-a407-0c3779dcf45c-webhook-cert\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.182156 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fj85\" (UniqueName: \"kubernetes.io/projected/d405d006-df42-4994-a407-0c3779dcf45c-kube-api-access-6fj85\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.182289 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d405d006-df42-4994-a407-0c3779dcf45c-apiservice-cert\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.188276 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d405d006-df42-4994-a407-0c3779dcf45c-webhook-cert\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.188751 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d405d006-df42-4994-a407-0c3779dcf45c-apiservice-cert\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.213943 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fj85\" (UniqueName: \"kubernetes.io/projected/d405d006-df42-4994-a407-0c3779dcf45c-kube-api-access-6fj85\") pod \"keystone-operator-controller-manager-5b7b95b7c4-9d5fb\" (UID: \"d405d006-df42-4994-a407-0c3779dcf45c\") " pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.386896 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:51 crc kubenswrapper[4692]: I0309 09:36:51.818858 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb"] Mar 09 09:36:51 crc kubenswrapper[4692]: W0309 09:36:51.827380 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd405d006_df42_4994_a407_0c3779dcf45c.slice/crio-ff36eec9c4f748055acb0e408510550270a240f1e4a43752e7e111392b212180 WatchSource:0}: Error finding container ff36eec9c4f748055acb0e408510550270a240f1e4a43752e7e111392b212180: Status 404 returned error can't find the container with id ff36eec9c4f748055acb0e408510550270a240f1e4a43752e7e111392b212180 Mar 09 09:36:52 crc kubenswrapper[4692]: I0309 09:36:52.779660 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" event={"ID":"d405d006-df42-4994-a407-0c3779dcf45c","Type":"ContainerStarted","Data":"ff36eec9c4f748055acb0e408510550270a240f1e4a43752e7e111392b212180"} Mar 09 09:36:55 crc kubenswrapper[4692]: I0309 09:36:55.805141 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" event={"ID":"d405d006-df42-4994-a407-0c3779dcf45c","Type":"ContainerStarted","Data":"2b6934cd6b3e0911aee2af834b17d9f07e4fcf6b6dfa9af36df5f38422f4d8a6"} Mar 09 09:36:55 crc kubenswrapper[4692]: I0309 09:36:55.805619 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:36:55 crc kubenswrapper[4692]: I0309 09:36:55.806807 4692 generic.go:334] "Generic (PLEG): container finished" podID="8075e838-fcea-4974-aead-79a1c9f94aa9" containerID="41e2107404f52cb893a332d456b57063181f111eb2bec62a9c373fa5055b5bdc" exitCode=0 Mar 09 09:36:55 crc kubenswrapper[4692]: I0309 09:36:55.806840 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"8075e838-fcea-4974-aead-79a1c9f94aa9","Type":"ContainerDied","Data":"41e2107404f52cb893a332d456b57063181f111eb2bec62a9c373fa5055b5bdc"} Mar 09 09:36:55 crc kubenswrapper[4692]: I0309 09:36:55.830245 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" podStartSLOduration=1.051526252 podStartE2EDuration="4.830225529s" podCreationTimestamp="2026-03-09 09:36:51 +0000 UTC" firstStartedPulling="2026-03-09 09:36:51.829900559 +0000 UTC m=+1012.654636140" lastFinishedPulling="2026-03-09 09:36:55.608599836 +0000 UTC m=+1016.433335417" observedRunningTime="2026-03-09 09:36:55.825902175 +0000 UTC m=+1016.650637776" watchObservedRunningTime="2026-03-09 09:36:55.830225529 +0000 UTC m=+1016.654961110" Mar 09 09:36:56 crc kubenswrapper[4692]: I0309 09:36:56.814626 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/rabbitmq-server-0" event={"ID":"8075e838-fcea-4974-aead-79a1c9f94aa9","Type":"ContainerStarted","Data":"8e687e7619c225e3ddd609a08bba88c92b367b21537267c304c8a0dd53097982"} Mar 09 09:36:56 crc kubenswrapper[4692]: I0309 09:36:56.815067 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:36:56 crc kubenswrapper[4692]: I0309 09:36:56.844438 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.938438535 podStartE2EDuration="41.844421601s" podCreationTimestamp="2026-03-09 09:36:15 +0000 UTC" firstStartedPulling="2026-03-09 09:36:17.491549589 +0000 UTC m=+978.316285170" lastFinishedPulling="2026-03-09 09:36:22.397532655 +0000 UTC m=+983.222268236" observedRunningTime="2026-03-09 09:36:56.840281793 +0000 UTC m=+1017.665017374" watchObservedRunningTime="2026-03-09 09:36:56.844421601 +0000 UTC m=+1017.669157182" Mar 09 09:37:01 crc kubenswrapper[4692]: I0309 09:37:01.130908 4692 scope.go:117] "RemoveContainer" containerID="00a653d5284093aa21c801feafebffd170508a7c8231e5b02ee93ae4526a2393" Mar 09 09:37:01 crc kubenswrapper[4692]: I0309 09:37:01.390976 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5b7b95b7c4-9d5fb" Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.030136 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/rabbitmq-server-0" Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.111096 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-index-p4x5b"] Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.111863 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-p4x5b" Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.114350 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-index-dockercfg-m6k2w" Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.120753 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-index-p4x5b"] Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.203147 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4zht\" (UniqueName: \"kubernetes.io/projected/e0288b74-65d6-41dd-9a84-b3953e84ba06-kube-api-access-r4zht\") pod \"barbican-operator-index-p4x5b\" (UID: \"e0288b74-65d6-41dd-9a84-b3953e84ba06\") " pod="openstack-operators/barbican-operator-index-p4x5b" Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.304588 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4zht\" (UniqueName: \"kubernetes.io/projected/e0288b74-65d6-41dd-9a84-b3953e84ba06-kube-api-access-r4zht\") pod \"barbican-operator-index-p4x5b\" (UID: \"e0288b74-65d6-41dd-9a84-b3953e84ba06\") " pod="openstack-operators/barbican-operator-index-p4x5b" Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.338401 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4zht\" (UniqueName: \"kubernetes.io/projected/e0288b74-65d6-41dd-9a84-b3953e84ba06-kube-api-access-r4zht\") pod \"barbican-operator-index-p4x5b\" (UID: \"e0288b74-65d6-41dd-9a84-b3953e84ba06\") " pod="openstack-operators/barbican-operator-index-p4x5b" Mar 09 09:37:07 crc kubenswrapper[4692]: I0309 09:37:07.440588 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-p4x5b" Mar 09 09:37:08 crc kubenswrapper[4692]: I0309 09:37:08.016226 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-index-p4x5b"] Mar 09 09:37:08 crc kubenswrapper[4692]: W0309 09:37:08.019657 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0288b74_65d6_41dd_9a84_b3953e84ba06.slice/crio-af97e231ac09e340db69f2a58b09db13e0fa9b76d24df228461f8eaf6512ecd8 WatchSource:0}: Error finding container af97e231ac09e340db69f2a58b09db13e0fa9b76d24df228461f8eaf6512ecd8: Status 404 returned error can't find the container with id af97e231ac09e340db69f2a58b09db13e0fa9b76d24df228461f8eaf6512ecd8 Mar 09 09:37:08 crc kubenswrapper[4692]: I0309 09:37:08.886686 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-p4x5b" event={"ID":"e0288b74-65d6-41dd-9a84-b3953e84ba06","Type":"ContainerStarted","Data":"761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb"} Mar 09 09:37:08 crc kubenswrapper[4692]: I0309 09:37:08.886988 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-p4x5b" event={"ID":"e0288b74-65d6-41dd-9a84-b3953e84ba06","Type":"ContainerStarted","Data":"af97e231ac09e340db69f2a58b09db13e0fa9b76d24df228461f8eaf6512ecd8"} Mar 09 09:37:08 crc kubenswrapper[4692]: I0309 09:37:08.901729 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-index-p4x5b" podStartSLOduration=1.243353026 podStartE2EDuration="1.901708788s" podCreationTimestamp="2026-03-09 09:37:07 +0000 UTC" firstStartedPulling="2026-03-09 09:37:08.022173897 +0000 UTC m=+1028.846909468" lastFinishedPulling="2026-03-09 09:37:08.680529649 +0000 UTC m=+1029.505265230" observedRunningTime="2026-03-09 09:37:08.897756416 +0000 UTC m=+1029.722491997" watchObservedRunningTime="2026-03-09 09:37:08.901708788 +0000 UTC m=+1029.726444379" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.277046 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/barbican-operator-index-p4x5b"] Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.277554 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/barbican-operator-index-p4x5b" podUID="e0288b74-65d6-41dd-9a84-b3953e84ba06" containerName="registry-server" containerID="cri-o://761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb" gracePeriod=2 Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.718022 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-p4x5b" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.861851 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4zht\" (UniqueName: \"kubernetes.io/projected/e0288b74-65d6-41dd-9a84-b3953e84ba06-kube-api-access-r4zht\") pod \"e0288b74-65d6-41dd-9a84-b3953e84ba06\" (UID: \"e0288b74-65d6-41dd-9a84-b3953e84ba06\") " Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.868326 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0288b74-65d6-41dd-9a84-b3953e84ba06-kube-api-access-r4zht" (OuterVolumeSpecName: "kube-api-access-r4zht") pod "e0288b74-65d6-41dd-9a84-b3953e84ba06" (UID: "e0288b74-65d6-41dd-9a84-b3953e84ba06"). InnerVolumeSpecName "kube-api-access-r4zht". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.892360 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-index-tjnxd"] Mar 09 09:37:11 crc kubenswrapper[4692]: E0309 09:37:11.892644 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0288b74-65d6-41dd-9a84-b3953e84ba06" containerName="registry-server" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.892666 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0288b74-65d6-41dd-9a84-b3953e84ba06" containerName="registry-server" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.892791 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0288b74-65d6-41dd-9a84-b3953e84ba06" containerName="registry-server" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.893291 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.906177 4692 generic.go:334] "Generic (PLEG): container finished" podID="e0288b74-65d6-41dd-9a84-b3953e84ba06" containerID="761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb" exitCode=0 Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.906382 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-p4x5b" event={"ID":"e0288b74-65d6-41dd-9a84-b3953e84ba06","Type":"ContainerDied","Data":"761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb"} Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.906414 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-p4x5b" event={"ID":"e0288b74-65d6-41dd-9a84-b3953e84ba06","Type":"ContainerDied","Data":"af97e231ac09e340db69f2a58b09db13e0fa9b76d24df228461f8eaf6512ecd8"} Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.906433 4692 scope.go:117] "RemoveContainer" containerID="761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.906560 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-p4x5b" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.908522 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-index-tjnxd"] Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.941677 4692 scope.go:117] "RemoveContainer" containerID="761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb" Mar 09 09:37:11 crc kubenswrapper[4692]: E0309 09:37:11.942085 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb\": container with ID starting with 761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb not found: ID does not exist" containerID="761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.942110 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb"} err="failed to get container status \"761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb\": rpc error: code = NotFound desc = could not find container \"761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb\": container with ID starting with 761cf6f82b6263b73305657863af80cfba35f17726f676917263e9137c8b33fb not found: ID does not exist" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.958666 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/barbican-operator-index-p4x5b"] Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.963730 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/barbican-operator-index-p4x5b"] Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.963912 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2rrk\" (UniqueName: \"kubernetes.io/projected/4da32689-47a4-449d-af7b-5eedecfe8e01-kube-api-access-c2rrk\") pod \"barbican-operator-index-tjnxd\" (UID: \"4da32689-47a4-449d-af7b-5eedecfe8e01\") " pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:11 crc kubenswrapper[4692]: I0309 09:37:11.964123 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4zht\" (UniqueName: \"kubernetes.io/projected/e0288b74-65d6-41dd-9a84-b3953e84ba06-kube-api-access-r4zht\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:12 crc kubenswrapper[4692]: I0309 09:37:12.064856 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2rrk\" (UniqueName: \"kubernetes.io/projected/4da32689-47a4-449d-af7b-5eedecfe8e01-kube-api-access-c2rrk\") pod \"barbican-operator-index-tjnxd\" (UID: \"4da32689-47a4-449d-af7b-5eedecfe8e01\") " pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:12 crc kubenswrapper[4692]: I0309 09:37:12.080682 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0288b74-65d6-41dd-9a84-b3953e84ba06" path="/var/lib/kubelet/pods/e0288b74-65d6-41dd-9a84-b3953e84ba06/volumes" Mar 09 09:37:12 crc kubenswrapper[4692]: I0309 09:37:12.084447 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2rrk\" (UniqueName: \"kubernetes.io/projected/4da32689-47a4-449d-af7b-5eedecfe8e01-kube-api-access-c2rrk\") pod \"barbican-operator-index-tjnxd\" (UID: \"4da32689-47a4-449d-af7b-5eedecfe8e01\") " pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:12 crc kubenswrapper[4692]: I0309 09:37:12.211413 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:12 crc kubenswrapper[4692]: I0309 09:37:12.616988 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-index-tjnxd"] Mar 09 09:37:12 crc kubenswrapper[4692]: I0309 09:37:12.913744 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-tjnxd" event={"ID":"4da32689-47a4-449d-af7b-5eedecfe8e01","Type":"ContainerStarted","Data":"746ab59db2183b57ab0a7f9420c02eb2e6e7eed7d5a44eed57c659efe59c5999"} Mar 09 09:37:14 crc kubenswrapper[4692]: I0309 09:37:14.934067 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-index-tjnxd" event={"ID":"4da32689-47a4-449d-af7b-5eedecfe8e01","Type":"ContainerStarted","Data":"a403ddd1642e0ab0e035bec019b80f4f9c88d2afefb193a5284ed7fc4e16f2cf"} Mar 09 09:37:14 crc kubenswrapper[4692]: I0309 09:37:14.951206 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-index-tjnxd" podStartSLOduration=2.526424911 podStartE2EDuration="3.951190286s" podCreationTimestamp="2026-03-09 09:37:11 +0000 UTC" firstStartedPulling="2026-03-09 09:37:12.625668395 +0000 UTC m=+1033.450403976" lastFinishedPulling="2026-03-09 09:37:14.05043377 +0000 UTC m=+1034.875169351" observedRunningTime="2026-03-09 09:37:14.948345905 +0000 UTC m=+1035.773081476" watchObservedRunningTime="2026-03-09 09:37:14.951190286 +0000 UTC m=+1035.775925877" Mar 09 09:37:22 crc kubenswrapper[4692]: I0309 09:37:22.212380 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:22 crc kubenswrapper[4692]: I0309 09:37:22.212822 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:22 crc kubenswrapper[4692]: I0309 09:37:22.237603 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:23 crc kubenswrapper[4692]: I0309 09:37:23.000002 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-index-tjnxd" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.247739 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-cb07-account-create-update-sknx7"] Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.249528 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.252664 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-db-create-md87c"] Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.253468 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.253891 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-db-secret" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.271680 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-cb07-account-create-update-sknx7"] Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.280974 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-create-md87c"] Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.358643 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-operator-scripts\") pod \"keystone-cb07-account-create-update-sknx7\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.358709 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p7xx\" (UniqueName: \"kubernetes.io/projected/c67531d0-eb39-4d97-aea1-4ff7c33fc191-kube-api-access-9p7xx\") pod \"keystone-db-create-md87c\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.358746 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c67531d0-eb39-4d97-aea1-4ff7c33fc191-operator-scripts\") pod \"keystone-db-create-md87c\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.358767 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g7l2\" (UniqueName: \"kubernetes.io/projected/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-kube-api-access-9g7l2\") pod \"keystone-cb07-account-create-update-sknx7\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.459850 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-operator-scripts\") pod \"keystone-cb07-account-create-update-sknx7\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.459918 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p7xx\" (UniqueName: \"kubernetes.io/projected/c67531d0-eb39-4d97-aea1-4ff7c33fc191-kube-api-access-9p7xx\") pod \"keystone-db-create-md87c\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.459969 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c67531d0-eb39-4d97-aea1-4ff7c33fc191-operator-scripts\") pod \"keystone-db-create-md87c\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.460006 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g7l2\" (UniqueName: \"kubernetes.io/projected/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-kube-api-access-9g7l2\") pod \"keystone-cb07-account-create-update-sknx7\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.460740 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-operator-scripts\") pod \"keystone-cb07-account-create-update-sknx7\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.460889 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c67531d0-eb39-4d97-aea1-4ff7c33fc191-operator-scripts\") pod \"keystone-db-create-md87c\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.479721 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p7xx\" (UniqueName: \"kubernetes.io/projected/c67531d0-eb39-4d97-aea1-4ff7c33fc191-kube-api-access-9p7xx\") pod \"keystone-db-create-md87c\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.479716 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g7l2\" (UniqueName: \"kubernetes.io/projected/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-kube-api-access-9g7l2\") pod \"keystone-cb07-account-create-update-sknx7\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.570328 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.579630 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:25 crc kubenswrapper[4692]: I0309 09:37:25.972898 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-cb07-account-create-update-sknx7"] Mar 09 09:37:26 crc kubenswrapper[4692]: I0309 09:37:26.006965 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" event={"ID":"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc","Type":"ContainerStarted","Data":"9ffdd7d827387774c8c5429c912018a3c3e1e529fa263eefb8e96cd88dac1fff"} Mar 09 09:37:26 crc kubenswrapper[4692]: I0309 09:37:26.079410 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-create-md87c"] Mar 09 09:37:26 crc kubenswrapper[4692]: W0309 09:37:26.093958 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc67531d0_eb39_4d97_aea1_4ff7c33fc191.slice/crio-a8076254664574528828429e6f1270826da7ff921800242213baaae6acb3e2ae WatchSource:0}: Error finding container a8076254664574528828429e6f1270826da7ff921800242213baaae6acb3e2ae: Status 404 returned error can't find the container with id a8076254664574528828429e6f1270826da7ff921800242213baaae6acb3e2ae Mar 09 09:37:27 crc kubenswrapper[4692]: I0309 09:37:27.013976 4692 generic.go:334] "Generic (PLEG): container finished" podID="e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc" containerID="87872a7b618c5191831e36a7519ff32436d180558935d34db116b477921f497c" exitCode=0 Mar 09 09:37:27 crc kubenswrapper[4692]: I0309 09:37:27.014027 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" event={"ID":"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc","Type":"ContainerDied","Data":"87872a7b618c5191831e36a7519ff32436d180558935d34db116b477921f497c"} Mar 09 09:37:27 crc kubenswrapper[4692]: I0309 09:37:27.015577 4692 generic.go:334] "Generic (PLEG): container finished" podID="c67531d0-eb39-4d97-aea1-4ff7c33fc191" containerID="536fe1e790da3e170a5692de5f9cf6ab816b621b035dca5db8bd6ecf2190efdc" exitCode=0 Mar 09 09:37:27 crc kubenswrapper[4692]: I0309 09:37:27.015613 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-create-md87c" event={"ID":"c67531d0-eb39-4d97-aea1-4ff7c33fc191","Type":"ContainerDied","Data":"536fe1e790da3e170a5692de5f9cf6ab816b621b035dca5db8bd6ecf2190efdc"} Mar 09 09:37:27 crc kubenswrapper[4692]: I0309 09:37:27.015655 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-create-md87c" event={"ID":"c67531d0-eb39-4d97-aea1-4ff7c33fc191","Type":"ContainerStarted","Data":"a8076254664574528828429e6f1270826da7ff921800242213baaae6acb3e2ae"} Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.434767 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.446639 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.505122 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p7xx\" (UniqueName: \"kubernetes.io/projected/c67531d0-eb39-4d97-aea1-4ff7c33fc191-kube-api-access-9p7xx\") pod \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.505260 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c67531d0-eb39-4d97-aea1-4ff7c33fc191-operator-scripts\") pod \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\" (UID: \"c67531d0-eb39-4d97-aea1-4ff7c33fc191\") " Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.505464 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g7l2\" (UniqueName: \"kubernetes.io/projected/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-kube-api-access-9g7l2\") pod \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.505500 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-operator-scripts\") pod \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\" (UID: \"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc\") " Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.506023 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c67531d0-eb39-4d97-aea1-4ff7c33fc191-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c67531d0-eb39-4d97-aea1-4ff7c33fc191" (UID: "c67531d0-eb39-4d97-aea1-4ff7c33fc191"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.506660 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc" (UID: "e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.511030 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-kube-api-access-9g7l2" (OuterVolumeSpecName: "kube-api-access-9g7l2") pod "e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc" (UID: "e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc"). InnerVolumeSpecName "kube-api-access-9g7l2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.511080 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67531d0-eb39-4d97-aea1-4ff7c33fc191-kube-api-access-9p7xx" (OuterVolumeSpecName: "kube-api-access-9p7xx") pod "c67531d0-eb39-4d97-aea1-4ff7c33fc191" (UID: "c67531d0-eb39-4d97-aea1-4ff7c33fc191"). InnerVolumeSpecName "kube-api-access-9p7xx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.606836 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p7xx\" (UniqueName: \"kubernetes.io/projected/c67531d0-eb39-4d97-aea1-4ff7c33fc191-kube-api-access-9p7xx\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.606872 4692 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c67531d0-eb39-4d97-aea1-4ff7c33fc191-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.606881 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g7l2\" (UniqueName: \"kubernetes.io/projected/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-kube-api-access-9g7l2\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:28 crc kubenswrapper[4692]: I0309 09:37:28.606889 4692 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:29 crc kubenswrapper[4692]: I0309 09:37:29.028424 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-create-md87c" Mar 09 09:37:29 crc kubenswrapper[4692]: I0309 09:37:29.028421 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-create-md87c" event={"ID":"c67531d0-eb39-4d97-aea1-4ff7c33fc191","Type":"ContainerDied","Data":"a8076254664574528828429e6f1270826da7ff921800242213baaae6acb3e2ae"} Mar 09 09:37:29 crc kubenswrapper[4692]: I0309 09:37:29.028854 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8076254664574528828429e6f1270826da7ff921800242213baaae6acb3e2ae" Mar 09 09:37:29 crc kubenswrapper[4692]: I0309 09:37:29.030666 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" event={"ID":"e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc","Type":"ContainerDied","Data":"9ffdd7d827387774c8c5429c912018a3c3e1e529fa263eefb8e96cd88dac1fff"} Mar 09 09:37:29 crc kubenswrapper[4692]: I0309 09:37:29.030703 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ffdd7d827387774c8c5429c912018a3c3e1e529fa263eefb8e96cd88dac1fff" Mar 09 09:37:29 crc kubenswrapper[4692]: I0309 09:37:29.030739 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cb07-account-create-update-sknx7" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.801135 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-db-sync-l8bgh"] Mar 09 09:37:30 crc kubenswrapper[4692]: E0309 09:37:30.801660 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc" containerName="mariadb-account-create-update" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.801673 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc" containerName="mariadb-account-create-update" Mar 09 09:37:30 crc kubenswrapper[4692]: E0309 09:37:30.801695 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67531d0-eb39-4d97-aea1-4ff7c33fc191" containerName="mariadb-database-create" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.801702 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67531d0-eb39-4d97-aea1-4ff7c33fc191" containerName="mariadb-database-create" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.801800 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc" containerName="mariadb-account-create-update" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.801808 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67531d0-eb39-4d97-aea1-4ff7c33fc191" containerName="mariadb-database-create" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.802236 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.803790 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-config-data" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.803828 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-scripts" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.803895 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-keystone-dockercfg-dmcpw" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.803968 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.810625 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-l8bgh"] Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.937136 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-config-data\") pod \"keystone-db-sync-l8bgh\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:30 crc kubenswrapper[4692]: I0309 09:37:30.937376 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bflx5\" (UniqueName: \"kubernetes.io/projected/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-kube-api-access-bflx5\") pod \"keystone-db-sync-l8bgh\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:31 crc kubenswrapper[4692]: I0309 09:37:31.038247 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bflx5\" (UniqueName: \"kubernetes.io/projected/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-kube-api-access-bflx5\") pod \"keystone-db-sync-l8bgh\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:31 crc kubenswrapper[4692]: I0309 09:37:31.038357 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-config-data\") pod \"keystone-db-sync-l8bgh\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:31 crc kubenswrapper[4692]: I0309 09:37:31.044142 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-config-data\") pod \"keystone-db-sync-l8bgh\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:31 crc kubenswrapper[4692]: I0309 09:37:31.055360 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bflx5\" (UniqueName: \"kubernetes.io/projected/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-kube-api-access-bflx5\") pod \"keystone-db-sync-l8bgh\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:31 crc kubenswrapper[4692]: I0309 09:37:31.117636 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:31 crc kubenswrapper[4692]: I0309 09:37:31.497455 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-l8bgh"] Mar 09 09:37:31 crc kubenswrapper[4692]: W0309 09:37:31.506135 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba1c1067_d220_48ff_b60c_d0947ee5b1d9.slice/crio-eb765c1d73bed534fb680462c8c1fec786f40d4cd01a909df8601767a4159b9a WatchSource:0}: Error finding container eb765c1d73bed534fb680462c8c1fec786f40d4cd01a909df8601767a4159b9a: Status 404 returned error can't find the container with id eb765c1d73bed534fb680462c8c1fec786f40d4cd01a909df8601767a4159b9a Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.048313 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" event={"ID":"ba1c1067-d220-48ff-b60c-d0947ee5b1d9","Type":"ContainerStarted","Data":"eb765c1d73bed534fb680462c8c1fec786f40d4cd01a909df8601767a4159b9a"} Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.131092 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw"] Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.132838 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.141957 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw"] Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.144801 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lthqw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.253271 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-util\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.253361 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmdst\" (UniqueName: \"kubernetes.io/projected/8901d82d-d48e-4b38-b746-78df5bf40e7d-kube-api-access-qmdst\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.253440 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-bundle\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.355005 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmdst\" (UniqueName: \"kubernetes.io/projected/8901d82d-d48e-4b38-b746-78df5bf40e7d-kube-api-access-qmdst\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.355071 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-bundle\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.355126 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-util\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.355683 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-util\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.355820 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-bundle\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.373017 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmdst\" (UniqueName: \"kubernetes.io/projected/8901d82d-d48e-4b38-b746-78df5bf40e7d-kube-api-access-qmdst\") pod \"30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.462508 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:32 crc kubenswrapper[4692]: I0309 09:37:32.900235 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw"] Mar 09 09:37:32 crc kubenswrapper[4692]: W0309 09:37:32.915983 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8901d82d_d48e_4b38_b746_78df5bf40e7d.slice/crio-250feba960d4cfa7d5c31221f3112cee46911b9bd4cab1b89c08af3b98a83381 WatchSource:0}: Error finding container 250feba960d4cfa7d5c31221f3112cee46911b9bd4cab1b89c08af3b98a83381: Status 404 returned error can't find the container with id 250feba960d4cfa7d5c31221f3112cee46911b9bd4cab1b89c08af3b98a83381 Mar 09 09:37:33 crc kubenswrapper[4692]: I0309 09:37:33.056397 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" event={"ID":"8901d82d-d48e-4b38-b746-78df5bf40e7d","Type":"ContainerStarted","Data":"250feba960d4cfa7d5c31221f3112cee46911b9bd4cab1b89c08af3b98a83381"} Mar 09 09:37:34 crc kubenswrapper[4692]: I0309 09:37:34.072080 4692 generic.go:334] "Generic (PLEG): container finished" podID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerID="b009b3b27630e7b0846ab70bc0e47211c6e6929de37676ce7151e1d2ebe7c02b" exitCode=0 Mar 09 09:37:34 crc kubenswrapper[4692]: I0309 09:37:34.084331 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" event={"ID":"8901d82d-d48e-4b38-b746-78df5bf40e7d","Type":"ContainerDied","Data":"b009b3b27630e7b0846ab70bc0e47211c6e6929de37676ce7151e1d2ebe7c02b"} Mar 09 09:37:40 crc kubenswrapper[4692]: I0309 09:37:40.754882 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" event={"ID":"ba1c1067-d220-48ff-b60c-d0947ee5b1d9","Type":"ContainerStarted","Data":"7ea425a2b3f008a6d3f1042de8b2947a69718e271a3a4348a6c56f9e8fa2f5b8"} Mar 09 09:37:40 crc kubenswrapper[4692]: I0309 09:37:40.756633 4692 generic.go:334] "Generic (PLEG): container finished" podID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerID="e491de2d075aeb80494e618ea002b1161e8b5561242ef7bb70b8954126861794" exitCode=0 Mar 09 09:37:40 crc kubenswrapper[4692]: I0309 09:37:40.756664 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" event={"ID":"8901d82d-d48e-4b38-b746-78df5bf40e7d","Type":"ContainerDied","Data":"e491de2d075aeb80494e618ea002b1161e8b5561242ef7bb70b8954126861794"} Mar 09 09:37:40 crc kubenswrapper[4692]: I0309 09:37:40.781715 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" podStartSLOduration=2.031078313 podStartE2EDuration="10.781693961s" podCreationTimestamp="2026-03-09 09:37:30 +0000 UTC" firstStartedPulling="2026-03-09 09:37:31.50884464 +0000 UTC m=+1052.333580221" lastFinishedPulling="2026-03-09 09:37:40.259460288 +0000 UTC m=+1061.084195869" observedRunningTime="2026-03-09 09:37:40.774488966 +0000 UTC m=+1061.599224557" watchObservedRunningTime="2026-03-09 09:37:40.781693961 +0000 UTC m=+1061.606429552" Mar 09 09:37:41 crc kubenswrapper[4692]: I0309 09:37:41.763961 4692 generic.go:334] "Generic (PLEG): container finished" podID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerID="c48b0c83954f3be5df9cbc9cfe2aa4f5fa38a03fdd19d6580da7120287aec5bd" exitCode=0 Mar 09 09:37:41 crc kubenswrapper[4692]: I0309 09:37:41.764002 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" event={"ID":"8901d82d-d48e-4b38-b746-78df5bf40e7d","Type":"ContainerDied","Data":"c48b0c83954f3be5df9cbc9cfe2aa4f5fa38a03fdd19d6580da7120287aec5bd"} Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.150937 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.241645 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-util\") pod \"8901d82d-d48e-4b38-b746-78df5bf40e7d\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.241691 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-bundle\") pod \"8901d82d-d48e-4b38-b746-78df5bf40e7d\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.241732 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmdst\" (UniqueName: \"kubernetes.io/projected/8901d82d-d48e-4b38-b746-78df5bf40e7d-kube-api-access-qmdst\") pod \"8901d82d-d48e-4b38-b746-78df5bf40e7d\" (UID: \"8901d82d-d48e-4b38-b746-78df5bf40e7d\") " Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.242714 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-bundle" (OuterVolumeSpecName: "bundle") pod "8901d82d-d48e-4b38-b746-78df5bf40e7d" (UID: "8901d82d-d48e-4b38-b746-78df5bf40e7d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.243773 4692 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.252009 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-util" (OuterVolumeSpecName: "util") pod "8901d82d-d48e-4b38-b746-78df5bf40e7d" (UID: "8901d82d-d48e-4b38-b746-78df5bf40e7d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.252150 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8901d82d-d48e-4b38-b746-78df5bf40e7d-kube-api-access-qmdst" (OuterVolumeSpecName: "kube-api-access-qmdst") pod "8901d82d-d48e-4b38-b746-78df5bf40e7d" (UID: "8901d82d-d48e-4b38-b746-78df5bf40e7d"). InnerVolumeSpecName "kube-api-access-qmdst". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.345088 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmdst\" (UniqueName: \"kubernetes.io/projected/8901d82d-d48e-4b38-b746-78df5bf40e7d-kube-api-access-qmdst\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.345117 4692 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8901d82d-d48e-4b38-b746-78df5bf40e7d-util\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.779126 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" event={"ID":"8901d82d-d48e-4b38-b746-78df5bf40e7d","Type":"ContainerDied","Data":"250feba960d4cfa7d5c31221f3112cee46911b9bd4cab1b89c08af3b98a83381"} Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.779193 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="250feba960d4cfa7d5c31221f3112cee46911b9bd4cab1b89c08af3b98a83381" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.779580 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw" Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.780512 4692 generic.go:334] "Generic (PLEG): container finished" podID="ba1c1067-d220-48ff-b60c-d0947ee5b1d9" containerID="7ea425a2b3f008a6d3f1042de8b2947a69718e271a3a4348a6c56f9e8fa2f5b8" exitCode=0 Mar 09 09:37:43 crc kubenswrapper[4692]: I0309 09:37:43.780599 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" event={"ID":"ba1c1067-d220-48ff-b60c-d0947ee5b1d9","Type":"ContainerDied","Data":"7ea425a2b3f008a6d3f1042de8b2947a69718e271a3a4348a6c56f9e8fa2f5b8"} Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.041558 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.176032 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bflx5\" (UniqueName: \"kubernetes.io/projected/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-kube-api-access-bflx5\") pod \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.176102 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-config-data\") pod \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\" (UID: \"ba1c1067-d220-48ff-b60c-d0947ee5b1d9\") " Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.181231 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-kube-api-access-bflx5" (OuterVolumeSpecName: "kube-api-access-bflx5") pod "ba1c1067-d220-48ff-b60c-d0947ee5b1d9" (UID: "ba1c1067-d220-48ff-b60c-d0947ee5b1d9"). InnerVolumeSpecName "kube-api-access-bflx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.210111 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-config-data" (OuterVolumeSpecName: "config-data") pod "ba1c1067-d220-48ff-b60c-d0947ee5b1d9" (UID: "ba1c1067-d220-48ff-b60c-d0947ee5b1d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.277570 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bflx5\" (UniqueName: \"kubernetes.io/projected/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-kube-api-access-bflx5\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.277605 4692 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1c1067-d220-48ff-b60c-d0947ee5b1d9-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.798345 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" event={"ID":"ba1c1067-d220-48ff-b60c-d0947ee5b1d9","Type":"ContainerDied","Data":"eb765c1d73bed534fb680462c8c1fec786f40d4cd01a909df8601767a4159b9a"} Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.798405 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb765c1d73bed534fb680462c8c1fec786f40d4cd01a909df8601767a4159b9a" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.798485 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-db-sync-l8bgh" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.987044 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-pg4dm"] Mar 09 09:37:45 crc kubenswrapper[4692]: E0309 09:37:45.987400 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1c1067-d220-48ff-b60c-d0947ee5b1d9" containerName="keystone-db-sync" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.987423 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1c1067-d220-48ff-b60c-d0947ee5b1d9" containerName="keystone-db-sync" Mar 09 09:37:45 crc kubenswrapper[4692]: E0309 09:37:45.987435 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerName="pull" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.987443 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerName="pull" Mar 09 09:37:45 crc kubenswrapper[4692]: E0309 09:37:45.987456 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerName="extract" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.987464 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerName="extract" Mar 09 09:37:45 crc kubenswrapper[4692]: E0309 09:37:45.987480 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerName="util" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.987487 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerName="util" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.987649 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="8901d82d-d48e-4b38-b746-78df5bf40e7d" containerName="extract" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.987681 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1c1067-d220-48ff-b60c-d0947ee5b1d9" containerName="keystone-db-sync" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.988258 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.990248 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-config-data" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.990385 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.990411 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-scripts" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.993410 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-keystone-dockercfg-dmcpw" Mar 09 09:37:45 crc kubenswrapper[4692]: I0309 09:37:45.993941 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"osp-secret" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.006437 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-pg4dm"] Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.089184 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-scripts\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.089231 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-fernet-keys\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.089261 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-credential-keys\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.089328 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdvv9\" (UniqueName: \"kubernetes.io/projected/f420a643-bfec-4568-8785-b1b18c1cf320-kube-api-access-vdvv9\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.089365 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-config-data\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.190958 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-fernet-keys\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.191026 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-credential-keys\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.191084 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdvv9\" (UniqueName: \"kubernetes.io/projected/f420a643-bfec-4568-8785-b1b18c1cf320-kube-api-access-vdvv9\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.191108 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-config-data\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.191213 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-scripts\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.200765 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-scripts\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.201204 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-credential-keys\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.201349 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-config-data\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.201391 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-fernet-keys\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.207689 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdvv9\" (UniqueName: \"kubernetes.io/projected/f420a643-bfec-4568-8785-b1b18c1cf320-kube-api-access-vdvv9\") pod \"keystone-bootstrap-pg4dm\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.306922 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.572233 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-pg4dm"] Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.807283 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" event={"ID":"f420a643-bfec-4568-8785-b1b18c1cf320","Type":"ContainerStarted","Data":"cc28cb1a5e009ae73ba2700723be66b48e4d67e1c1d76931d82c683f007bfd38"} Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.807611 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" event={"ID":"f420a643-bfec-4568-8785-b1b18c1cf320","Type":"ContainerStarted","Data":"1a76762bfea80c7402ae20ce34bbf2e73d84f3814e545116d3a51fb68555d006"} Mar 09 09:37:46 crc kubenswrapper[4692]: I0309 09:37:46.822955 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" podStartSLOduration=1.822937517 podStartE2EDuration="1.822937517s" podCreationTimestamp="2026-03-09 09:37:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:37:46.822210707 +0000 UTC m=+1067.646946318" watchObservedRunningTime="2026-03-09 09:37:46.822937517 +0000 UTC m=+1067.647673098" Mar 09 09:37:49 crc kubenswrapper[4692]: I0309 09:37:49.828825 4692 generic.go:334] "Generic (PLEG): container finished" podID="f420a643-bfec-4568-8785-b1b18c1cf320" containerID="cc28cb1a5e009ae73ba2700723be66b48e4d67e1c1d76931d82c683f007bfd38" exitCode=0 Mar 09 09:37:49 crc kubenswrapper[4692]: I0309 09:37:49.828903 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" event={"ID":"f420a643-bfec-4568-8785-b1b18c1cf320","Type":"ContainerDied","Data":"cc28cb1a5e009ae73ba2700723be66b48e4d67e1c1d76931d82c683f007bfd38"} Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.082837 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.162846 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-scripts\") pod \"f420a643-bfec-4568-8785-b1b18c1cf320\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.162915 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdvv9\" (UniqueName: \"kubernetes.io/projected/f420a643-bfec-4568-8785-b1b18c1cf320-kube-api-access-vdvv9\") pod \"f420a643-bfec-4568-8785-b1b18c1cf320\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.162972 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-config-data\") pod \"f420a643-bfec-4568-8785-b1b18c1cf320\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.163018 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-fernet-keys\") pod \"f420a643-bfec-4568-8785-b1b18c1cf320\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.163049 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-credential-keys\") pod \"f420a643-bfec-4568-8785-b1b18c1cf320\" (UID: \"f420a643-bfec-4568-8785-b1b18c1cf320\") " Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.168591 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f420a643-bfec-4568-8785-b1b18c1cf320" (UID: "f420a643-bfec-4568-8785-b1b18c1cf320"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.168771 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f420a643-bfec-4568-8785-b1b18c1cf320" (UID: "f420a643-bfec-4568-8785-b1b18c1cf320"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.168787 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f420a643-bfec-4568-8785-b1b18c1cf320-kube-api-access-vdvv9" (OuterVolumeSpecName: "kube-api-access-vdvv9") pod "f420a643-bfec-4568-8785-b1b18c1cf320" (UID: "f420a643-bfec-4568-8785-b1b18c1cf320"). InnerVolumeSpecName "kube-api-access-vdvv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.169282 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-scripts" (OuterVolumeSpecName: "scripts") pod "f420a643-bfec-4568-8785-b1b18c1cf320" (UID: "f420a643-bfec-4568-8785-b1b18c1cf320"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.182029 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-config-data" (OuterVolumeSpecName: "config-data") pod "f420a643-bfec-4568-8785-b1b18c1cf320" (UID: "f420a643-bfec-4568-8785-b1b18c1cf320"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.264141 4692 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.264196 4692 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.264207 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.264216 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdvv9\" (UniqueName: \"kubernetes.io/projected/f420a643-bfec-4568-8785-b1b18c1cf320-kube-api-access-vdvv9\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.264227 4692 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f420a643-bfec-4568-8785-b1b18c1cf320-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.841744 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" event={"ID":"f420a643-bfec-4568-8785-b1b18c1cf320","Type":"ContainerDied","Data":"1a76762bfea80c7402ae20ce34bbf2e73d84f3814e545116d3a51fb68555d006"} Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.841785 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-bootstrap-pg4dm" Mar 09 09:37:51 crc kubenswrapper[4692]: I0309 09:37:51.841788 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a76762bfea80c7402ae20ce34bbf2e73d84f3814e545116d3a51fb68555d006" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.006352 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-88b4f4fb5-8bz88"] Mar 09 09:37:52 crc kubenswrapper[4692]: E0309 09:37:52.006594 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f420a643-bfec-4568-8785-b1b18c1cf320" containerName="keystone-bootstrap" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.006606 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f420a643-bfec-4568-8785-b1b18c1cf320" containerName="keystone-bootstrap" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.006723 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f420a643-bfec-4568-8785-b1b18c1cf320" containerName="keystone-bootstrap" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.007110 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.010439 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-keystone-dockercfg-dmcpw" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.010484 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-scripts" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.010544 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone-config-data" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.010681 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"keystone" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.017693 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-88b4f4fb5-8bz88"] Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.178925 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-credential-keys\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.179305 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-scripts\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.179363 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-fernet-keys\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.179388 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp5kx\" (UniqueName: \"kubernetes.io/projected/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-kube-api-access-wp5kx\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.179422 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-config-data\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.269746 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x"] Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.270813 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.273438 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-7bqdv" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.273955 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-service-cert" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.280717 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-fernet-keys\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.280764 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp5kx\" (UniqueName: \"kubernetes.io/projected/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-kube-api-access-wp5kx\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.280792 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-config-data\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.280890 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-credential-keys\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.280933 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-scripts\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.284683 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-credential-keys\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.284833 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-scripts\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.288394 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x"] Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.290307 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-config-data\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.291288 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-fernet-keys\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.303686 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp5kx\" (UniqueName: \"kubernetes.io/projected/33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36-kube-api-access-wp5kx\") pod \"keystone-88b4f4fb5-8bz88\" (UID: \"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36\") " pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.326789 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.384704 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp4q7\" (UniqueName: \"kubernetes.io/projected/9f91a180-3832-448a-9744-efdb956c9bd0-kube-api-access-vp4q7\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.385070 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f91a180-3832-448a-9744-efdb956c9bd0-webhook-cert\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.385302 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f91a180-3832-448a-9744-efdb956c9bd0-apiservice-cert\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.486155 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp4q7\" (UniqueName: \"kubernetes.io/projected/9f91a180-3832-448a-9744-efdb956c9bd0-kube-api-access-vp4q7\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.486628 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f91a180-3832-448a-9744-efdb956c9bd0-webhook-cert\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.486670 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f91a180-3832-448a-9744-efdb956c9bd0-apiservice-cert\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.494368 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f91a180-3832-448a-9744-efdb956c9bd0-webhook-cert\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.511092 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f91a180-3832-448a-9744-efdb956c9bd0-apiservice-cert\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.518649 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp4q7\" (UniqueName: \"kubernetes.io/projected/9f91a180-3832-448a-9744-efdb956c9bd0-kube-api-access-vp4q7\") pod \"barbican-operator-controller-manager-64d5945c66-k4v8x\" (UID: \"9f91a180-3832-448a-9744-efdb956c9bd0\") " pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.641933 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.734548 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-88b4f4fb5-8bz88"] Mar 09 09:37:52 crc kubenswrapper[4692]: I0309 09:37:52.893404 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" event={"ID":"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36","Type":"ContainerStarted","Data":"ca18896567f383ca7b2c408a39b5a45528f8440d28700a72ae2979e6bfa29c40"} Mar 09 09:37:53 crc kubenswrapper[4692]: I0309 09:37:53.116286 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x"] Mar 09 09:37:53 crc kubenswrapper[4692]: W0309 09:37:53.118540 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f91a180_3832_448a_9744_efdb956c9bd0.slice/crio-b7ecddf732f255614d807747b28669d2f46500b024e9c6bf1a9d5a3e4431b2e8 WatchSource:0}: Error finding container b7ecddf732f255614d807747b28669d2f46500b024e9c6bf1a9d5a3e4431b2e8: Status 404 returned error can't find the container with id b7ecddf732f255614d807747b28669d2f46500b024e9c6bf1a9d5a3e4431b2e8 Mar 09 09:37:53 crc kubenswrapper[4692]: I0309 09:37:53.901463 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" event={"ID":"9f91a180-3832-448a-9744-efdb956c9bd0","Type":"ContainerStarted","Data":"b7ecddf732f255614d807747b28669d2f46500b024e9c6bf1a9d5a3e4431b2e8"} Mar 09 09:37:53 crc kubenswrapper[4692]: I0309 09:37:53.903398 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" event={"ID":"33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36","Type":"ContainerStarted","Data":"56c768e94bfc3d6b6226cf5ba136547508984808b903071b2c855241f2c90444"} Mar 09 09:37:53 crc kubenswrapper[4692]: I0309 09:37:53.903488 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:37:53 crc kubenswrapper[4692]: I0309 09:37:53.924203 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" podStartSLOduration=2.924186669 podStartE2EDuration="2.924186669s" podCreationTimestamp="2026-03-09 09:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:37:53.917824767 +0000 UTC m=+1074.742560368" watchObservedRunningTime="2026-03-09 09:37:53.924186669 +0000 UTC m=+1074.748922250" Mar 09 09:37:55 crc kubenswrapper[4692]: I0309 09:37:55.916066 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" event={"ID":"9f91a180-3832-448a-9744-efdb956c9bd0","Type":"ContainerStarted","Data":"40faad8338fa0ef68871683417a416b4fdbaadd8211d5c2ad6558127580dce2b"} Mar 09 09:37:55 crc kubenswrapper[4692]: I0309 09:37:55.916608 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:37:55 crc kubenswrapper[4692]: I0309 09:37:55.941790 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" podStartSLOduration=1.732230651 podStartE2EDuration="3.941766005s" podCreationTimestamp="2026-03-09 09:37:52 +0000 UTC" firstStartedPulling="2026-03-09 09:37:53.12090975 +0000 UTC m=+1073.945645331" lastFinishedPulling="2026-03-09 09:37:55.330445104 +0000 UTC m=+1076.155180685" observedRunningTime="2026-03-09 09:37:55.934882479 +0000 UTC m=+1076.759618070" watchObservedRunningTime="2026-03-09 09:37:55.941766005 +0000 UTC m=+1076.766501596" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.136741 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550818-7xlfm"] Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.138333 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550818-7xlfm" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.142163 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.142877 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.143111 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.148728 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550818-7xlfm"] Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.304733 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hvgr\" (UniqueName: \"kubernetes.io/projected/3e5bbd9f-03a6-419e-b495-df023635b82e-kube-api-access-2hvgr\") pod \"auto-csr-approver-29550818-7xlfm\" (UID: \"3e5bbd9f-03a6-419e-b495-df023635b82e\") " pod="openshift-infra/auto-csr-approver-29550818-7xlfm" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.406903 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hvgr\" (UniqueName: \"kubernetes.io/projected/3e5bbd9f-03a6-419e-b495-df023635b82e-kube-api-access-2hvgr\") pod \"auto-csr-approver-29550818-7xlfm\" (UID: \"3e5bbd9f-03a6-419e-b495-df023635b82e\") " pod="openshift-infra/auto-csr-approver-29550818-7xlfm" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.427548 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hvgr\" (UniqueName: \"kubernetes.io/projected/3e5bbd9f-03a6-419e-b495-df023635b82e-kube-api-access-2hvgr\") pod \"auto-csr-approver-29550818-7xlfm\" (UID: \"3e5bbd9f-03a6-419e-b495-df023635b82e\") " pod="openshift-infra/auto-csr-approver-29550818-7xlfm" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.458604 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550818-7xlfm" Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.904176 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550818-7xlfm"] Mar 09 09:38:00 crc kubenswrapper[4692]: W0309 09:38:00.913758 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e5bbd9f_03a6_419e_b495_df023635b82e.slice/crio-f055ea3e7840c8ef2307b1c18d72c7bcd5b75c1897a9516c9b31b0aec91dbd1a WatchSource:0}: Error finding container f055ea3e7840c8ef2307b1c18d72c7bcd5b75c1897a9516c9b31b0aec91dbd1a: Status 404 returned error can't find the container with id f055ea3e7840c8ef2307b1c18d72c7bcd5b75c1897a9516c9b31b0aec91dbd1a Mar 09 09:38:00 crc kubenswrapper[4692]: I0309 09:38:00.951406 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550818-7xlfm" event={"ID":"3e5bbd9f-03a6-419e-b495-df023635b82e","Type":"ContainerStarted","Data":"f055ea3e7840c8ef2307b1c18d72c7bcd5b75c1897a9516c9b31b0aec91dbd1a"} Mar 09 09:38:02 crc kubenswrapper[4692]: I0309 09:38:02.668471 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64d5945c66-k4v8x" Mar 09 09:38:03 crc kubenswrapper[4692]: I0309 09:38:03.972549 4692 generic.go:334] "Generic (PLEG): container finished" podID="3e5bbd9f-03a6-419e-b495-df023635b82e" containerID="f7f92cd200436cefbfd944307e7af99d76ac4b846f0aad05045381c2845036e5" exitCode=0 Mar 09 09:38:03 crc kubenswrapper[4692]: I0309 09:38:03.972612 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550818-7xlfm" event={"ID":"3e5bbd9f-03a6-419e-b495-df023635b82e","Type":"ContainerDied","Data":"f7f92cd200436cefbfd944307e7af99d76ac4b846f0aad05045381c2845036e5"} Mar 09 09:38:05 crc kubenswrapper[4692]: I0309 09:38:05.386389 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550818-7xlfm" Mar 09 09:38:05 crc kubenswrapper[4692]: I0309 09:38:05.483253 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hvgr\" (UniqueName: \"kubernetes.io/projected/3e5bbd9f-03a6-419e-b495-df023635b82e-kube-api-access-2hvgr\") pod \"3e5bbd9f-03a6-419e-b495-df023635b82e\" (UID: \"3e5bbd9f-03a6-419e-b495-df023635b82e\") " Mar 09 09:38:05 crc kubenswrapper[4692]: I0309 09:38:05.488577 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e5bbd9f-03a6-419e-b495-df023635b82e-kube-api-access-2hvgr" (OuterVolumeSpecName: "kube-api-access-2hvgr") pod "3e5bbd9f-03a6-419e-b495-df023635b82e" (UID: "3e5bbd9f-03a6-419e-b495-df023635b82e"). InnerVolumeSpecName "kube-api-access-2hvgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:38:05 crc kubenswrapper[4692]: I0309 09:38:05.584608 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hvgr\" (UniqueName: \"kubernetes.io/projected/3e5bbd9f-03a6-419e-b495-df023635b82e-kube-api-access-2hvgr\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:05 crc kubenswrapper[4692]: I0309 09:38:05.985369 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550818-7xlfm" event={"ID":"3e5bbd9f-03a6-419e-b495-df023635b82e","Type":"ContainerDied","Data":"f055ea3e7840c8ef2307b1c18d72c7bcd5b75c1897a9516c9b31b0aec91dbd1a"} Mar 09 09:38:05 crc kubenswrapper[4692]: I0309 09:38:05.985409 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f055ea3e7840c8ef2307b1c18d72c7bcd5b75c1897a9516c9b31b0aec91dbd1a" Mar 09 09:38:05 crc kubenswrapper[4692]: I0309 09:38:05.985415 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550818-7xlfm" Mar 09 09:38:06 crc kubenswrapper[4692]: I0309 09:38:06.450435 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550812-prhwl"] Mar 09 09:38:06 crc kubenswrapper[4692]: I0309 09:38:06.456441 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550812-prhwl"] Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.680852 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-mlnfz"] Mar 09 09:38:07 crc kubenswrapper[4692]: E0309 09:38:07.681151 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e5bbd9f-03a6-419e-b495-df023635b82e" containerName="oc" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.681190 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e5bbd9f-03a6-419e-b495-df023635b82e" containerName="oc" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.681310 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e5bbd9f-03a6-419e-b495-df023635b82e" containerName="oc" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.681805 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-mlnfz" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.684356 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-g7hdr" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.692297 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-mlnfz"] Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.813279 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z7rr\" (UniqueName: \"kubernetes.io/projected/960b5d9b-1f1e-4975-abf2-9cc1e84f20e0-kube-api-access-9z7rr\") pod \"swift-operator-index-mlnfz\" (UID: \"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0\") " pod="openstack-operators/swift-operator-index-mlnfz" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.915038 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z7rr\" (UniqueName: \"kubernetes.io/projected/960b5d9b-1f1e-4975-abf2-9cc1e84f20e0-kube-api-access-9z7rr\") pod \"swift-operator-index-mlnfz\" (UID: \"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0\") " pod="openstack-operators/swift-operator-index-mlnfz" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.932101 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z7rr\" (UniqueName: \"kubernetes.io/projected/960b5d9b-1f1e-4975-abf2-9cc1e84f20e0-kube-api-access-9z7rr\") pod \"swift-operator-index-mlnfz\" (UID: \"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0\") " pod="openstack-operators/swift-operator-index-mlnfz" Mar 09 09:38:07 crc kubenswrapper[4692]: I0309 09:38:07.997482 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-mlnfz" Mar 09 09:38:08 crc kubenswrapper[4692]: I0309 09:38:08.080889 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcaa35a6-212d-4f51-bdc6-77d45a0d589c" path="/var/lib/kubelet/pods/dcaa35a6-212d-4f51-bdc6-77d45a0d589c/volumes" Mar 09 09:38:08 crc kubenswrapper[4692]: I0309 09:38:08.544459 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-mlnfz"] Mar 09 09:38:08 crc kubenswrapper[4692]: W0309 09:38:08.551377 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod960b5d9b_1f1e_4975_abf2_9cc1e84f20e0.slice/crio-105f23dc4bf8c6e935d7c46fc3736fb3de69cd56e41957426664744b4378bf2f WatchSource:0}: Error finding container 105f23dc4bf8c6e935d7c46fc3736fb3de69cd56e41957426664744b4378bf2f: Status 404 returned error can't find the container with id 105f23dc4bf8c6e935d7c46fc3736fb3de69cd56e41957426664744b4378bf2f Mar 09 09:38:09 crc kubenswrapper[4692]: I0309 09:38:09.012297 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-mlnfz" event={"ID":"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0","Type":"ContainerStarted","Data":"105f23dc4bf8c6e935d7c46fc3736fb3de69cd56e41957426664744b4378bf2f"} Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.038048 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-mlnfz" event={"ID":"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0","Type":"ContainerStarted","Data":"bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7"} Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.057728 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-mlnfz" podStartSLOduration=2.170687449 podStartE2EDuration="5.057705901s" podCreationTimestamp="2026-03-09 09:38:07 +0000 UTC" firstStartedPulling="2026-03-09 09:38:08.553748286 +0000 UTC m=+1089.378483867" lastFinishedPulling="2026-03-09 09:38:11.440766728 +0000 UTC m=+1092.265502319" observedRunningTime="2026-03-09 09:38:12.051473273 +0000 UTC m=+1092.876208874" watchObservedRunningTime="2026-03-09 09:38:12.057705901 +0000 UTC m=+1092.882441492" Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.080429 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-mlnfz"] Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.685081 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-2mmxr"] Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.685909 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.694399 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-2mmxr"] Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.785806 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbnld\" (UniqueName: \"kubernetes.io/projected/31360b66-7736-41cf-bbff-ff901238bd28-kube-api-access-wbnld\") pod \"swift-operator-index-2mmxr\" (UID: \"31360b66-7736-41cf-bbff-ff901238bd28\") " pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.887137 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbnld\" (UniqueName: \"kubernetes.io/projected/31360b66-7736-41cf-bbff-ff901238bd28-kube-api-access-wbnld\") pod \"swift-operator-index-2mmxr\" (UID: \"31360b66-7736-41cf-bbff-ff901238bd28\") " pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:12 crc kubenswrapper[4692]: I0309 09:38:12.907929 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbnld\" (UniqueName: \"kubernetes.io/projected/31360b66-7736-41cf-bbff-ff901238bd28-kube-api-access-wbnld\") pod \"swift-operator-index-2mmxr\" (UID: \"31360b66-7736-41cf-bbff-ff901238bd28\") " pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:13 crc kubenswrapper[4692]: I0309 09:38:13.008149 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:13 crc kubenswrapper[4692]: I0309 09:38:13.422027 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-2mmxr"] Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.055990 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-2mmxr" event={"ID":"31360b66-7736-41cf-bbff-ff901238bd28","Type":"ContainerStarted","Data":"441252a0cc5e222b409e2afc9c00734235f059accb1229467178f3c1eb3218f3"} Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.056545 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-2mmxr" event={"ID":"31360b66-7736-41cf-bbff-ff901238bd28","Type":"ContainerStarted","Data":"943ef8c38e7eb4844b0273cc67c15dd11cf16b020a1b0a48b64600d1f3edaf1b"} Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.056217 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/swift-operator-index-mlnfz" podUID="960b5d9b-1f1e-4975-abf2-9cc1e84f20e0" containerName="registry-server" containerID="cri-o://bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7" gracePeriod=2 Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.076389 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-2mmxr" podStartSLOduration=2.028311988 podStartE2EDuration="2.076371086s" podCreationTimestamp="2026-03-09 09:38:12 +0000 UTC" firstStartedPulling="2026-03-09 09:38:13.432009343 +0000 UTC m=+1094.256744914" lastFinishedPulling="2026-03-09 09:38:13.480068431 +0000 UTC m=+1094.304804012" observedRunningTime="2026-03-09 09:38:14.075813371 +0000 UTC m=+1094.900548972" watchObservedRunningTime="2026-03-09 09:38:14.076371086 +0000 UTC m=+1094.901106667" Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.444177 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-mlnfz" Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.609990 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z7rr\" (UniqueName: \"kubernetes.io/projected/960b5d9b-1f1e-4975-abf2-9cc1e84f20e0-kube-api-access-9z7rr\") pod \"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0\" (UID: \"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0\") " Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.615277 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/960b5d9b-1f1e-4975-abf2-9cc1e84f20e0-kube-api-access-9z7rr" (OuterVolumeSpecName: "kube-api-access-9z7rr") pod "960b5d9b-1f1e-4975-abf2-9cc1e84f20e0" (UID: "960b5d9b-1f1e-4975-abf2-9cc1e84f20e0"). InnerVolumeSpecName "kube-api-access-9z7rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:38:14 crc kubenswrapper[4692]: I0309 09:38:14.711289 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9z7rr\" (UniqueName: \"kubernetes.io/projected/960b5d9b-1f1e-4975-abf2-9cc1e84f20e0-kube-api-access-9z7rr\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.093290 4692 generic.go:334] "Generic (PLEG): container finished" podID="960b5d9b-1f1e-4975-abf2-9cc1e84f20e0" containerID="bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7" exitCode=0 Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.093365 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-mlnfz" Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.093403 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-mlnfz" event={"ID":"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0","Type":"ContainerDied","Data":"bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7"} Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.093474 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-mlnfz" event={"ID":"960b5d9b-1f1e-4975-abf2-9cc1e84f20e0","Type":"ContainerDied","Data":"105f23dc4bf8c6e935d7c46fc3736fb3de69cd56e41957426664744b4378bf2f"} Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.093499 4692 scope.go:117] "RemoveContainer" containerID="bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7" Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.128880 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/swift-operator-index-mlnfz"] Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.140336 4692 scope.go:117] "RemoveContainer" containerID="bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7" Mar 09 09:38:15 crc kubenswrapper[4692]: E0309 09:38:15.140863 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7\": container with ID starting with bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7 not found: ID does not exist" containerID="bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7" Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.140908 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7"} err="failed to get container status \"bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7\": rpc error: code = NotFound desc = could not find container \"bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7\": container with ID starting with bd441b9bffb08a4fbf6636396bc6a54b3a6969a1a3ee7833ebb2be066f91afe7 not found: ID does not exist" Mar 09 09:38:15 crc kubenswrapper[4692]: I0309 09:38:15.145702 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/swift-operator-index-mlnfz"] Mar 09 09:38:16 crc kubenswrapper[4692]: I0309 09:38:16.079550 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="960b5d9b-1f1e-4975-abf2-9cc1e84f20e0" path="/var/lib/kubelet/pods/960b5d9b-1f1e-4975-abf2-9cc1e84f20e0/volumes" Mar 09 09:38:23 crc kubenswrapper[4692]: I0309 09:38:23.008550 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:23 crc kubenswrapper[4692]: I0309 09:38:23.009145 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:23 crc kubenswrapper[4692]: I0309 09:38:23.038122 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:23 crc kubenswrapper[4692]: I0309 09:38:23.170945 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-2mmxr" Mar 09 09:38:23 crc kubenswrapper[4692]: I0309 09:38:23.987625 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/keystone-88b4f4fb5-8bz88" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.451115 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-990a-account-create-update-kcj5x"] Mar 09 09:38:25 crc kubenswrapper[4692]: E0309 09:38:25.451555 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="960b5d9b-1f1e-4975-abf2-9cc1e84f20e0" containerName="registry-server" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.451572 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="960b5d9b-1f1e-4975-abf2-9cc1e84f20e0" containerName="registry-server" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.451722 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="960b5d9b-1f1e-4975-abf2-9cc1e84f20e0" containerName="registry-server" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.452291 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.455301 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-db-secret" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.457008 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-db-create-z4kvk"] Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.458059 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.469900 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-990a-account-create-update-kcj5x"] Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.477753 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-create-z4kvk"] Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.557253 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-operator-scripts\") pod \"barbican-990a-account-create-update-kcj5x\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.557325 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgvgk\" (UniqueName: \"kubernetes.io/projected/d85ecf61-be76-4439-a305-d4a64fae30f3-kube-api-access-wgvgk\") pod \"barbican-db-create-z4kvk\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.557363 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d85ecf61-be76-4439-a305-d4a64fae30f3-operator-scripts\") pod \"barbican-db-create-z4kvk\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.557386 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt6nx\" (UniqueName: \"kubernetes.io/projected/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-kube-api-access-kt6nx\") pod \"barbican-990a-account-create-update-kcj5x\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.658520 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d85ecf61-be76-4439-a305-d4a64fae30f3-operator-scripts\") pod \"barbican-db-create-z4kvk\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.658574 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt6nx\" (UniqueName: \"kubernetes.io/projected/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-kube-api-access-kt6nx\") pod \"barbican-990a-account-create-update-kcj5x\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.658631 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-operator-scripts\") pod \"barbican-990a-account-create-update-kcj5x\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.658672 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgvgk\" (UniqueName: \"kubernetes.io/projected/d85ecf61-be76-4439-a305-d4a64fae30f3-kube-api-access-wgvgk\") pod \"barbican-db-create-z4kvk\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.659278 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d85ecf61-be76-4439-a305-d4a64fae30f3-operator-scripts\") pod \"barbican-db-create-z4kvk\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.659709 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-operator-scripts\") pod \"barbican-990a-account-create-update-kcj5x\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.679616 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt6nx\" (UniqueName: \"kubernetes.io/projected/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-kube-api-access-kt6nx\") pod \"barbican-990a-account-create-update-kcj5x\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.685317 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgvgk\" (UniqueName: \"kubernetes.io/projected/d85ecf61-be76-4439-a305-d4a64fae30f3-kube-api-access-wgvgk\") pod \"barbican-db-create-z4kvk\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.772039 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:25 crc kubenswrapper[4692]: I0309 09:38:25.784351 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.312101 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-990a-account-create-update-kcj5x"] Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.390736 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-create-z4kvk"] Mar 09 09:38:26 crc kubenswrapper[4692]: W0309 09:38:26.405230 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd85ecf61_be76_4439_a305_d4a64fae30f3.slice/crio-272b994cf853f025d487ac56dfa2139e3483efd6948a37cf39808e1abd7fcc9a WatchSource:0}: Error finding container 272b994cf853f025d487ac56dfa2139e3483efd6948a37cf39808e1abd7fcc9a: Status 404 returned error can't find the container with id 272b994cf853f025d487ac56dfa2139e3483efd6948a37cf39808e1abd7fcc9a Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.720769 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x"] Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.722020 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.723803 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-lthqw" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.729850 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x"] Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.777624 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-bundle\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.777754 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-util\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.777830 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmvh9\" (UniqueName: \"kubernetes.io/projected/4f6faf73-b691-41b4-a0d7-f653bc0587c6-kube-api-access-cmvh9\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.878981 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-bundle\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.879132 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-util\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.879565 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-bundle\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.879597 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-util\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.879708 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmvh9\" (UniqueName: \"kubernetes.io/projected/4f6faf73-b691-41b4-a0d7-f653bc0587c6-kube-api-access-cmvh9\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:26 crc kubenswrapper[4692]: I0309 09:38:26.899263 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmvh9\" (UniqueName: \"kubernetes.io/projected/4f6faf73-b691-41b4-a0d7-f653bc0587c6-kube-api-access-cmvh9\") pod \"fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.090664 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.177802 4692 generic.go:334] "Generic (PLEG): container finished" podID="d85ecf61-be76-4439-a305-d4a64fae30f3" containerID="8468b2f345979089bbe533ee7e8ea32f7e7d3806c32bb748fae2a1772cdd78b8" exitCode=0 Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.177874 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-create-z4kvk" event={"ID":"d85ecf61-be76-4439-a305-d4a64fae30f3","Type":"ContainerDied","Data":"8468b2f345979089bbe533ee7e8ea32f7e7d3806c32bb748fae2a1772cdd78b8"} Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.177906 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-create-z4kvk" event={"ID":"d85ecf61-be76-4439-a305-d4a64fae30f3","Type":"ContainerStarted","Data":"272b994cf853f025d487ac56dfa2139e3483efd6948a37cf39808e1abd7fcc9a"} Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.183601 4692 generic.go:334] "Generic (PLEG): container finished" podID="6ff25a5a-862c-4367-a86d-3f6a5afec5a7" containerID="b6ee2239f970169d9a9900d51f829435dd2ed84e3c45b62a56a77b685c8e6700" exitCode=0 Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.183637 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" event={"ID":"6ff25a5a-862c-4367-a86d-3f6a5afec5a7","Type":"ContainerDied","Data":"b6ee2239f970169d9a9900d51f829435dd2ed84e3c45b62a56a77b685c8e6700"} Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.183661 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" event={"ID":"6ff25a5a-862c-4367-a86d-3f6a5afec5a7","Type":"ContainerStarted","Data":"758d02d3aa236f8decaca5c1fa40725cce1623f602357a98f5bb8420aa7dc946"} Mar 09 09:38:27 crc kubenswrapper[4692]: I0309 09:38:27.521975 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x"] Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.192215 4692 generic.go:334] "Generic (PLEG): container finished" podID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerID="267e04b403f1ce3a08c38e8043c19421cf023da6ebd00aeb6b81ec9a7a64118c" exitCode=0 Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.192343 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" event={"ID":"4f6faf73-b691-41b4-a0d7-f653bc0587c6","Type":"ContainerDied","Data":"267e04b403f1ce3a08c38e8043c19421cf023da6ebd00aeb6b81ec9a7a64118c"} Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.192623 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" event={"ID":"4f6faf73-b691-41b4-a0d7-f653bc0587c6","Type":"ContainerStarted","Data":"d2901a1186adfa807020d8b0de02933b2ec9417a47ed297d83a171790b4892c8"} Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.193933 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.713337 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.718336 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.803642 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt6nx\" (UniqueName: \"kubernetes.io/projected/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-kube-api-access-kt6nx\") pod \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.803814 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgvgk\" (UniqueName: \"kubernetes.io/projected/d85ecf61-be76-4439-a305-d4a64fae30f3-kube-api-access-wgvgk\") pod \"d85ecf61-be76-4439-a305-d4a64fae30f3\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.803842 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-operator-scripts\") pod \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\" (UID: \"6ff25a5a-862c-4367-a86d-3f6a5afec5a7\") " Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.803876 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d85ecf61-be76-4439-a305-d4a64fae30f3-operator-scripts\") pod \"d85ecf61-be76-4439-a305-d4a64fae30f3\" (UID: \"d85ecf61-be76-4439-a305-d4a64fae30f3\") " Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.804473 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d85ecf61-be76-4439-a305-d4a64fae30f3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d85ecf61-be76-4439-a305-d4a64fae30f3" (UID: "d85ecf61-be76-4439-a305-d4a64fae30f3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.804485 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ff25a5a-862c-4367-a86d-3f6a5afec5a7" (UID: "6ff25a5a-862c-4367-a86d-3f6a5afec5a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.804692 4692 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.804715 4692 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d85ecf61-be76-4439-a305-d4a64fae30f3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.808860 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-kube-api-access-kt6nx" (OuterVolumeSpecName: "kube-api-access-kt6nx") pod "6ff25a5a-862c-4367-a86d-3f6a5afec5a7" (UID: "6ff25a5a-862c-4367-a86d-3f6a5afec5a7"). InnerVolumeSpecName "kube-api-access-kt6nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.812319 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d85ecf61-be76-4439-a305-d4a64fae30f3-kube-api-access-wgvgk" (OuterVolumeSpecName: "kube-api-access-wgvgk") pod "d85ecf61-be76-4439-a305-d4a64fae30f3" (UID: "d85ecf61-be76-4439-a305-d4a64fae30f3"). InnerVolumeSpecName "kube-api-access-wgvgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.905780 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgvgk\" (UniqueName: \"kubernetes.io/projected/d85ecf61-be76-4439-a305-d4a64fae30f3-kube-api-access-wgvgk\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:28 crc kubenswrapper[4692]: I0309 09:38:28.905817 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt6nx\" (UniqueName: \"kubernetes.io/projected/6ff25a5a-862c-4367-a86d-3f6a5afec5a7-kube-api-access-kt6nx\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.200365 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" event={"ID":"6ff25a5a-862c-4367-a86d-3f6a5afec5a7","Type":"ContainerDied","Data":"758d02d3aa236f8decaca5c1fa40725cce1623f602357a98f5bb8420aa7dc946"} Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.200394 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-990a-account-create-update-kcj5x" Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.200401 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="758d02d3aa236f8decaca5c1fa40725cce1623f602357a98f5bb8420aa7dc946" Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.202071 4692 generic.go:334] "Generic (PLEG): container finished" podID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerID="72194919815aed9ea2a16457ec2503503dc1a127fadc2a3fdf4ac6a4f759110a" exitCode=0 Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.202125 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" event={"ID":"4f6faf73-b691-41b4-a0d7-f653bc0587c6","Type":"ContainerDied","Data":"72194919815aed9ea2a16457ec2503503dc1a127fadc2a3fdf4ac6a4f759110a"} Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.204932 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-create-z4kvk" event={"ID":"d85ecf61-be76-4439-a305-d4a64fae30f3","Type":"ContainerDied","Data":"272b994cf853f025d487ac56dfa2139e3483efd6948a37cf39808e1abd7fcc9a"} Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.204965 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="272b994cf853f025d487ac56dfa2139e3483efd6948a37cf39808e1abd7fcc9a" Mar 09 09:38:29 crc kubenswrapper[4692]: I0309 09:38:29.204974 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-create-z4kvk" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.218502 4692 generic.go:334] "Generic (PLEG): container finished" podID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerID="fca94b34870fa7fc9f09272b03a23da0aa0f687d6297e5a360a5423dd42b401f" exitCode=0 Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.218821 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" event={"ID":"4f6faf73-b691-41b4-a0d7-f653bc0587c6","Type":"ContainerDied","Data":"fca94b34870fa7fc9f09272b03a23da0aa0f687d6297e5a360a5423dd42b401f"} Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.728417 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-db-sync-b22vl"] Mar 09 09:38:30 crc kubenswrapper[4692]: E0309 09:38:30.728676 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ff25a5a-862c-4367-a86d-3f6a5afec5a7" containerName="mariadb-account-create-update" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.728692 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ff25a5a-862c-4367-a86d-3f6a5afec5a7" containerName="mariadb-account-create-update" Mar 09 09:38:30 crc kubenswrapper[4692]: E0309 09:38:30.728706 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d85ecf61-be76-4439-a305-d4a64fae30f3" containerName="mariadb-database-create" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.728714 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d85ecf61-be76-4439-a305-d4a64fae30f3" containerName="mariadb-database-create" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.728833 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d85ecf61-be76-4439-a305-d4a64fae30f3" containerName="mariadb-database-create" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.728847 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ff25a5a-862c-4367-a86d-3f6a5afec5a7" containerName="mariadb-account-create-update" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.729264 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.730926 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-barbican-dockercfg-xg6pw" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.731105 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-config-data" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.740076 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-sync-b22vl"] Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.829191 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/189d4c9b-7202-4625-a552-fa66691a4bb7-db-sync-config-data\") pod \"barbican-db-sync-b22vl\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.829310 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wf7k\" (UniqueName: \"kubernetes.io/projected/189d4c9b-7202-4625-a552-fa66691a4bb7-kube-api-access-4wf7k\") pod \"barbican-db-sync-b22vl\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.930310 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/189d4c9b-7202-4625-a552-fa66691a4bb7-db-sync-config-data\") pod \"barbican-db-sync-b22vl\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.930373 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wf7k\" (UniqueName: \"kubernetes.io/projected/189d4c9b-7202-4625-a552-fa66691a4bb7-kube-api-access-4wf7k\") pod \"barbican-db-sync-b22vl\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.942757 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/189d4c9b-7202-4625-a552-fa66691a4bb7-db-sync-config-data\") pod \"barbican-db-sync-b22vl\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:30 crc kubenswrapper[4692]: I0309 09:38:30.952002 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wf7k\" (UniqueName: \"kubernetes.io/projected/189d4c9b-7202-4625-a552-fa66691a4bb7-kube-api-access-4wf7k\") pod \"barbican-db-sync-b22vl\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.092330 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.513135 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-db-sync-b22vl"] Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.556651 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.641900 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmvh9\" (UniqueName: \"kubernetes.io/projected/4f6faf73-b691-41b4-a0d7-f653bc0587c6-kube-api-access-cmvh9\") pod \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.642777 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-util\") pod \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.642940 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-bundle\") pod \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\" (UID: \"4f6faf73-b691-41b4-a0d7-f653bc0587c6\") " Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.644132 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-bundle" (OuterVolumeSpecName: "bundle") pod "4f6faf73-b691-41b4-a0d7-f653bc0587c6" (UID: "4f6faf73-b691-41b4-a0d7-f653bc0587c6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.647126 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f6faf73-b691-41b4-a0d7-f653bc0587c6-kube-api-access-cmvh9" (OuterVolumeSpecName: "kube-api-access-cmvh9") pod "4f6faf73-b691-41b4-a0d7-f653bc0587c6" (UID: "4f6faf73-b691-41b4-a0d7-f653bc0587c6"). InnerVolumeSpecName "kube-api-access-cmvh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.661088 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-util" (OuterVolumeSpecName: "util") pod "4f6faf73-b691-41b4-a0d7-f653bc0587c6" (UID: "4f6faf73-b691-41b4-a0d7-f653bc0587c6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.744668 4692 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-bundle\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.744719 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmvh9\" (UniqueName: \"kubernetes.io/projected/4f6faf73-b691-41b4-a0d7-f653bc0587c6-kube-api-access-cmvh9\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:31 crc kubenswrapper[4692]: I0309 09:38:31.744732 4692 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4f6faf73-b691-41b4-a0d7-f653bc0587c6-util\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:32 crc kubenswrapper[4692]: I0309 09:38:32.240828 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" event={"ID":"4f6faf73-b691-41b4-a0d7-f653bc0587c6","Type":"ContainerDied","Data":"d2901a1186adfa807020d8b0de02933b2ec9417a47ed297d83a171790b4892c8"} Mar 09 09:38:32 crc kubenswrapper[4692]: I0309 09:38:32.240878 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2901a1186adfa807020d8b0de02933b2ec9417a47ed297d83a171790b4892c8" Mar 09 09:38:32 crc kubenswrapper[4692]: I0309 09:38:32.240838 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x" Mar 09 09:38:32 crc kubenswrapper[4692]: I0309 09:38:32.242781 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-b22vl" event={"ID":"189d4c9b-7202-4625-a552-fa66691a4bb7","Type":"ContainerStarted","Data":"6c3c858a82b65439cbe23e8cbb9e094e94781382cc1776717727ac9f4511cdb8"} Mar 09 09:38:37 crc kubenswrapper[4692]: I0309 09:38:37.279552 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-b22vl" event={"ID":"189d4c9b-7202-4625-a552-fa66691a4bb7","Type":"ContainerStarted","Data":"bcf9e6dabd32349786cd17cb89ac2cb3686f1a8f80e2fce363be6ac585c8c326"} Mar 09 09:38:37 crc kubenswrapper[4692]: I0309 09:38:37.304605 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-db-sync-b22vl" podStartSLOduration=2.339131269 podStartE2EDuration="7.304590611s" podCreationTimestamp="2026-03-09 09:38:30 +0000 UTC" firstStartedPulling="2026-03-09 09:38:31.520379532 +0000 UTC m=+1112.345115113" lastFinishedPulling="2026-03-09 09:38:36.485838884 +0000 UTC m=+1117.310574455" observedRunningTime="2026-03-09 09:38:37.299965126 +0000 UTC m=+1118.124700707" watchObservedRunningTime="2026-03-09 09:38:37.304590611 +0000 UTC m=+1118.129326192" Mar 09 09:38:40 crc kubenswrapper[4692]: I0309 09:38:40.300028 4692 generic.go:334] "Generic (PLEG): container finished" podID="189d4c9b-7202-4625-a552-fa66691a4bb7" containerID="bcf9e6dabd32349786cd17cb89ac2cb3686f1a8f80e2fce363be6ac585c8c326" exitCode=0 Mar 09 09:38:40 crc kubenswrapper[4692]: I0309 09:38:40.300123 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-b22vl" event={"ID":"189d4c9b-7202-4625-a552-fa66691a4bb7","Type":"ContainerDied","Data":"bcf9e6dabd32349786cd17cb89ac2cb3686f1a8f80e2fce363be6ac585c8c326"} Mar 09 09:38:41 crc kubenswrapper[4692]: I0309 09:38:41.718159 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:41 crc kubenswrapper[4692]: I0309 09:38:41.874990 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wf7k\" (UniqueName: \"kubernetes.io/projected/189d4c9b-7202-4625-a552-fa66691a4bb7-kube-api-access-4wf7k\") pod \"189d4c9b-7202-4625-a552-fa66691a4bb7\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " Mar 09 09:38:41 crc kubenswrapper[4692]: I0309 09:38:41.875055 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/189d4c9b-7202-4625-a552-fa66691a4bb7-db-sync-config-data\") pod \"189d4c9b-7202-4625-a552-fa66691a4bb7\" (UID: \"189d4c9b-7202-4625-a552-fa66691a4bb7\") " Mar 09 09:38:41 crc kubenswrapper[4692]: I0309 09:38:41.889361 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/189d4c9b-7202-4625-a552-fa66691a4bb7-kube-api-access-4wf7k" (OuterVolumeSpecName: "kube-api-access-4wf7k") pod "189d4c9b-7202-4625-a552-fa66691a4bb7" (UID: "189d4c9b-7202-4625-a552-fa66691a4bb7"). InnerVolumeSpecName "kube-api-access-4wf7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:38:41 crc kubenswrapper[4692]: I0309 09:38:41.897092 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/189d4c9b-7202-4625-a552-fa66691a4bb7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "189d4c9b-7202-4625-a552-fa66691a4bb7" (UID: "189d4c9b-7202-4625-a552-fa66691a4bb7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:38:41 crc kubenswrapper[4692]: I0309 09:38:41.976690 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wf7k\" (UniqueName: \"kubernetes.io/projected/189d4c9b-7202-4625-a552-fa66691a4bb7-kube-api-access-4wf7k\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:41 crc kubenswrapper[4692]: I0309 09:38:41.976734 4692 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/189d4c9b-7202-4625-a552-fa66691a4bb7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.312209 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-db-sync-b22vl" event={"ID":"189d4c9b-7202-4625-a552-fa66691a4bb7","Type":"ContainerDied","Data":"6c3c858a82b65439cbe23e8cbb9e094e94781382cc1776717727ac9f4511cdb8"} Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.312247 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c3c858a82b65439cbe23e8cbb9e094e94781382cc1776717727ac9f4511cdb8" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.312253 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-db-sync-b22vl" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.535725 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx"] Mar 09 09:38:42 crc kubenswrapper[4692]: E0309 09:38:42.538735 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerName="extract" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.538764 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerName="extract" Mar 09 09:38:42 crc kubenswrapper[4692]: E0309 09:38:42.538788 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerName="util" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.538794 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerName="util" Mar 09 09:38:42 crc kubenswrapper[4692]: E0309 09:38:42.538813 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerName="pull" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.538819 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerName="pull" Mar 09 09:38:42 crc kubenswrapper[4692]: E0309 09:38:42.538840 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="189d4c9b-7202-4625-a552-fa66691a4bb7" containerName="barbican-db-sync" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.538846 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="189d4c9b-7202-4625-a552-fa66691a4bb7" containerName="barbican-db-sync" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.539777 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f6faf73-b691-41b4-a0d7-f653bc0587c6" containerName="extract" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.539793 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="189d4c9b-7202-4625-a552-fa66691a4bb7" containerName="barbican-db-sync" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.541930 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.552086 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-worker-config-data" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.552116 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-barbican-dockercfg-xg6pw" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.552371 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-config-data" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.581232 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx"] Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.601295 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p"] Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.602421 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.609972 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-keystone-listener-config-data" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.619189 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p"] Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.666472 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg"] Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.667530 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.674198 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.674443 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-x59vp" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.681565 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg"] Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.685430 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klgpt\" (UniqueName: \"kubernetes.io/projected/2a94f29a-c39a-4c48-be28-472570beb2b3-kube-api-access-klgpt\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.685495 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a94f29a-c39a-4c48-be28-472570beb2b3-config-data-custom\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.685528 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a94f29a-c39a-4c48-be28-472570beb2b3-config-data\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.685640 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a94f29a-c39a-4c48-be28-472570beb2b3-logs\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.713723 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/barbican-api-5b4dcd568-j4snq"] Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.718747 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.721344 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-api-5b4dcd568-j4snq"] Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.722483 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"barbican-api-config-data" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787011 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8de0018-50da-4b21-a94c-ef673de69ab1-logs\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787047 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d6362adf-592b-479c-ab1f-6dfd352e0570-apiservice-cert\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787112 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t2ks\" (UniqueName: \"kubernetes.io/projected/d6362adf-592b-479c-ab1f-6dfd352e0570-kube-api-access-9t2ks\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787247 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a94f29a-c39a-4c48-be28-472570beb2b3-logs\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787299 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klgpt\" (UniqueName: \"kubernetes.io/projected/2a94f29a-c39a-4c48-be28-472570beb2b3-kube-api-access-klgpt\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787345 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a94f29a-c39a-4c48-be28-472570beb2b3-config-data-custom\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787374 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a94f29a-c39a-4c48-be28-472570beb2b3-config-data\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787400 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d6362adf-592b-479c-ab1f-6dfd352e0570-webhook-cert\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787466 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8de0018-50da-4b21-a94c-ef673de69ab1-config-data\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787487 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh5lk\" (UniqueName: \"kubernetes.io/projected/e8de0018-50da-4b21-a94c-ef673de69ab1-kube-api-access-kh5lk\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787536 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8de0018-50da-4b21-a94c-ef673de69ab1-config-data-custom\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.787641 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a94f29a-c39a-4c48-be28-472570beb2b3-logs\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.793749 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a94f29a-c39a-4c48-be28-472570beb2b3-config-data\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.811690 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a94f29a-c39a-4c48-be28-472570beb2b3-config-data-custom\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.811747 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klgpt\" (UniqueName: \"kubernetes.io/projected/2a94f29a-c39a-4c48-be28-472570beb2b3-kube-api-access-klgpt\") pod \"barbican-worker-786f5f954f-mmjsx\" (UID: \"2a94f29a-c39a-4c48-be28-472570beb2b3\") " pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888743 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d6362adf-592b-479c-ab1f-6dfd352e0570-webhook-cert\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888808 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8de0018-50da-4b21-a94c-ef673de69ab1-config-data\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888837 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-logs\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888869 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh5lk\" (UniqueName: \"kubernetes.io/projected/e8de0018-50da-4b21-a94c-ef673de69ab1-kube-api-access-kh5lk\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888904 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8de0018-50da-4b21-a94c-ef673de69ab1-config-data-custom\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888930 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8de0018-50da-4b21-a94c-ef673de69ab1-logs\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888949 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d6362adf-592b-479c-ab1f-6dfd352e0570-apiservice-cert\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.888972 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-config-data\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.889008 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-config-data-custom\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.889030 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t2ks\" (UniqueName: \"kubernetes.io/projected/d6362adf-592b-479c-ab1f-6dfd352e0570-kube-api-access-9t2ks\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.889066 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b65jm\" (UniqueName: \"kubernetes.io/projected/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-kube-api-access-b65jm\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.890452 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e8de0018-50da-4b21-a94c-ef673de69ab1-logs\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.892385 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d6362adf-592b-479c-ab1f-6dfd352e0570-webhook-cert\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.893722 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8de0018-50da-4b21-a94c-ef673de69ab1-config-data\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.893994 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e8de0018-50da-4b21-a94c-ef673de69ab1-config-data-custom\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.895400 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.897733 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d6362adf-592b-479c-ab1f-6dfd352e0570-apiservice-cert\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.917054 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t2ks\" (UniqueName: \"kubernetes.io/projected/d6362adf-592b-479c-ab1f-6dfd352e0570-kube-api-access-9t2ks\") pod \"swift-operator-controller-manager-6b7c9bbf4c-49qbg\" (UID: \"d6362adf-592b-479c-ab1f-6dfd352e0570\") " pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.927151 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh5lk\" (UniqueName: \"kubernetes.io/projected/e8de0018-50da-4b21-a94c-ef673de69ab1-kube-api-access-kh5lk\") pod \"barbican-keystone-listener-74f57bcf5d-nbr2p\" (UID: \"e8de0018-50da-4b21-a94c-ef673de69ab1\") " pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.991184 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-config-data-custom\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.991251 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b65jm\" (UniqueName: \"kubernetes.io/projected/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-kube-api-access-b65jm\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.992109 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.992241 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-logs\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.992293 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-config-data\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.992770 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-logs\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:42 crc kubenswrapper[4692]: I0309 09:38:42.997271 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-config-data-custom\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.002598 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-config-data\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.053844 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b65jm\" (UniqueName: \"kubernetes.io/projected/47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4-kube-api-access-b65jm\") pod \"barbican-api-5b4dcd568-j4snq\" (UID: \"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4\") " pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.224157 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.339291 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.430645 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx"] Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.643241 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg"] Mar 09 09:38:43 crc kubenswrapper[4692]: W0309 09:38:43.643883 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6362adf_592b_479c_ab1f_6dfd352e0570.slice/crio-2c4dda0df041db7e739a7d22ea493cffe5c483f9ac3efe98fbcbab0bd7824955 WatchSource:0}: Error finding container 2c4dda0df041db7e739a7d22ea493cffe5c483f9ac3efe98fbcbab0bd7824955: Status 404 returned error can't find the container with id 2c4dda0df041db7e739a7d22ea493cffe5c483f9ac3efe98fbcbab0bd7824955 Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.839898 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p"] Mar 09 09:38:43 crc kubenswrapper[4692]: W0309 09:38:43.847053 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8de0018_50da_4b21_a94c_ef673de69ab1.slice/crio-a2dc197764ed9866486a644a5e3e9793c52c7b09cc031e9788439bd4e1d6a6b4 WatchSource:0}: Error finding container a2dc197764ed9866486a644a5e3e9793c52c7b09cc031e9788439bd4e1d6a6b4: Status 404 returned error can't find the container with id a2dc197764ed9866486a644a5e3e9793c52c7b09cc031e9788439bd4e1d6a6b4 Mar 09 09:38:43 crc kubenswrapper[4692]: I0309 09:38:43.931609 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/barbican-api-5b4dcd568-j4snq"] Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.333682 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" event={"ID":"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4","Type":"ContainerStarted","Data":"b052a8a26d80d978ce67bfb80859d9596633c2fc7e05435c78fff0d36dccdc9c"} Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.333998 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.334016 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" event={"ID":"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4","Type":"ContainerStarted","Data":"02cc31460c19bc789182e4785e4d7a3b7d8d87fcf2b47b3e382a586736ab7c56"} Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.334027 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" event={"ID":"47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4","Type":"ContainerStarted","Data":"07b08195f024c9f49df150aa4027b3e162f884f71c8c086a16f512dbbd984750"} Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.335503 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" event={"ID":"d6362adf-592b-479c-ab1f-6dfd352e0570","Type":"ContainerStarted","Data":"2c4dda0df041db7e739a7d22ea493cffe5c483f9ac3efe98fbcbab0bd7824955"} Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.336858 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" event={"ID":"2a94f29a-c39a-4c48-be28-472570beb2b3","Type":"ContainerStarted","Data":"49540cfac8d443911657ae67385dad4d3fa302d272aa17fdb130577fe7b09fec"} Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.337936 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" event={"ID":"e8de0018-50da-4b21-a94c-ef673de69ab1","Type":"ContainerStarted","Data":"a2dc197764ed9866486a644a5e3e9793c52c7b09cc031e9788439bd4e1d6a6b4"} Mar 09 09:38:44 crc kubenswrapper[4692]: I0309 09:38:44.375739 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" podStartSLOduration=2.375720319 podStartE2EDuration="2.375720319s" podCreationTimestamp="2026-03-09 09:38:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:38:44.349263294 +0000 UTC m=+1125.173998905" watchObservedRunningTime="2026-03-09 09:38:44.375720319 +0000 UTC m=+1125.200455900" Mar 09 09:38:45 crc kubenswrapper[4692]: I0309 09:38:45.375193 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.390888 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" event={"ID":"d6362adf-592b-479c-ab1f-6dfd352e0570","Type":"ContainerStarted","Data":"54d825e29ac86435afdb98b18dc15f45278837d102bcf9c0dc58d07162313d63"} Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.391252 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.393498 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" event={"ID":"2a94f29a-c39a-4c48-be28-472570beb2b3","Type":"ContainerStarted","Data":"85208539c91c22932fbb4f5022e5c9c0a4a449420cfc17005b12e6235cb58942"} Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.393528 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" event={"ID":"2a94f29a-c39a-4c48-be28-472570beb2b3","Type":"ContainerStarted","Data":"af47ffa0217b4cb28783a3c1ee3a101c1311150daf3f321f5ac04832c04a4bf5"} Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.395139 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" event={"ID":"e8de0018-50da-4b21-a94c-ef673de69ab1","Type":"ContainerStarted","Data":"1df972914f8711684213c47ece17aa5aa317c5e2833d812fff1c0c4fcf302af4"} Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.395208 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" event={"ID":"e8de0018-50da-4b21-a94c-ef673de69ab1","Type":"ContainerStarted","Data":"dde9851dff1e8e10ee35c1a342a26c89649a617fa37586a0ee901968ba7c2d84"} Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.414902 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" podStartSLOduration=2.505733523 podStartE2EDuration="5.414880803s" podCreationTimestamp="2026-03-09 09:38:42 +0000 UTC" firstStartedPulling="2026-03-09 09:38:43.647032596 +0000 UTC m=+1124.471768177" lastFinishedPulling="2026-03-09 09:38:46.556179876 +0000 UTC m=+1127.380915457" observedRunningTime="2026-03-09 09:38:47.408408608 +0000 UTC m=+1128.233144209" watchObservedRunningTime="2026-03-09 09:38:47.414880803 +0000 UTC m=+1128.239616384" Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.432831 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-keystone-listener-74f57bcf5d-nbr2p" podStartSLOduration=2.7289298349999997 podStartE2EDuration="5.432811868s" podCreationTimestamp="2026-03-09 09:38:42 +0000 UTC" firstStartedPulling="2026-03-09 09:38:43.849169559 +0000 UTC m=+1124.673905140" lastFinishedPulling="2026-03-09 09:38:46.553051592 +0000 UTC m=+1127.377787173" observedRunningTime="2026-03-09 09:38:47.426351203 +0000 UTC m=+1128.251086794" watchObservedRunningTime="2026-03-09 09:38:47.432811868 +0000 UTC m=+1128.257547449" Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.455305 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/barbican-worker-786f5f954f-mmjsx" podStartSLOduration=2.680174978 podStartE2EDuration="5.455254894s" podCreationTimestamp="2026-03-09 09:38:42 +0000 UTC" firstStartedPulling="2026-03-09 09:38:43.45548655 +0000 UTC m=+1124.280222131" lastFinishedPulling="2026-03-09 09:38:46.230566466 +0000 UTC m=+1127.055302047" observedRunningTime="2026-03-09 09:38:47.44435268 +0000 UTC m=+1128.269088261" watchObservedRunningTime="2026-03-09 09:38:47.455254894 +0000 UTC m=+1128.279990475" Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.624086 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:38:47 crc kubenswrapper[4692]: I0309 09:38:47.624136 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:38:53 crc kubenswrapper[4692]: I0309 09:38:53.006643 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6b7c9bbf4c-49qbg" Mar 09 09:38:55 crc kubenswrapper[4692]: I0309 09:38:55.764543 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:55 crc kubenswrapper[4692]: I0309 09:38:55.766710 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/barbican-api-5b4dcd568-j4snq" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.510110 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.514642 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.517984 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-conf" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.518126 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-swift-dockercfg-qm65s" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.518223 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-storage-config-data" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.518974 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-files" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.537872 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.615899 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-z69jg"] Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.616867 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.618993 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.619149 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-proxy-config-data" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.619716 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.627061 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-lock\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.627126 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-cache\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.627153 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.627193 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.627305 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8hp6\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-kube-api-access-t8hp6\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.679243 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-z69jg"] Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728294 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-scripts\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728356 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-lock\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728388 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-cache\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728412 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728437 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728460 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-swiftconf\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728520 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8hp6\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-kube-api-access-t8hp6\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728551 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9csm\" (UniqueName: \"kubernetes.io/projected/1eb95f27-3506-4519-8401-35420bdbb7e6-kube-api-access-b9csm\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728596 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-dispersionconf\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728619 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-ring-data-devices\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728645 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eb95f27-3506-4519-8401-35420bdbb7e6-etc-swift\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728832 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") device mount path \"/mnt/openstack/pv12\"" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.728889 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-cache\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: E0309 09:38:56.728989 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:38:56 crc kubenswrapper[4692]: E0309 09:38:56.729016 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:38:56 crc kubenswrapper[4692]: E0309 09:38:56.729063 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift podName:243cb7f7-785c-4681-b344-ea4073562c84 nodeName:}" failed. No retries permitted until 2026-03-09 09:38:57.22904367 +0000 UTC m=+1138.053779261 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift") pod "swift-storage-0" (UID: "243cb7f7-785c-4681-b344-ea4073562c84") : configmap "swift-ring-files" not found Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.742917 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-lock\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.759810 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8hp6\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-kube-api-access-t8hp6\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.759887 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.829740 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-dispersionconf\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.829791 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-ring-data-devices\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.829823 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eb95f27-3506-4519-8401-35420bdbb7e6-etc-swift\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.829878 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-scripts\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.829920 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-swiftconf\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.829996 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9csm\" (UniqueName: \"kubernetes.io/projected/1eb95f27-3506-4519-8401-35420bdbb7e6-kube-api-access-b9csm\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.830785 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-ring-data-devices\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.830852 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-scripts\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.831296 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eb95f27-3506-4519-8401-35420bdbb7e6-etc-swift\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.835146 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-dispersionconf\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.835201 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-swiftconf\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.856797 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9csm\" (UniqueName: \"kubernetes.io/projected/1eb95f27-3506-4519-8401-35420bdbb7e6-kube-api-access-b9csm\") pod \"swift-ring-rebalance-z69jg\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:56 crc kubenswrapper[4692]: I0309 09:38:56.942268 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:38:57 crc kubenswrapper[4692]: I0309 09:38:57.237014 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:57 crc kubenswrapper[4692]: E0309 09:38:57.238293 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:38:57 crc kubenswrapper[4692]: E0309 09:38:57.238310 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:38:57 crc kubenswrapper[4692]: E0309 09:38:57.238343 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift podName:243cb7f7-785c-4681-b344-ea4073562c84 nodeName:}" failed. No retries permitted until 2026-03-09 09:38:58.238329982 +0000 UTC m=+1139.063065563 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift") pod "swift-storage-0" (UID: "243cb7f7-785c-4681-b344-ea4073562c84") : configmap "swift-ring-files" not found Mar 09 09:38:57 crc kubenswrapper[4692]: I0309 09:38:57.840347 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-z69jg"] Mar 09 09:38:57 crc kubenswrapper[4692]: W0309 09:38:57.864014 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1eb95f27_3506_4519_8401_35420bdbb7e6.slice/crio-42d8ea28f99a69b068612b9a9e6a1cb73d12e022997a3c2c44f71846e299dddc WatchSource:0}: Error finding container 42d8ea28f99a69b068612b9a9e6a1cb73d12e022997a3c2c44f71846e299dddc: Status 404 returned error can't find the container with id 42d8ea28f99a69b068612b9a9e6a1cb73d12e022997a3c2c44f71846e299dddc Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.312011 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:38:58 crc kubenswrapper[4692]: E0309 09:38:58.312173 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:38:58 crc kubenswrapper[4692]: E0309 09:38:58.312188 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:38:58 crc kubenswrapper[4692]: E0309 09:38:58.312238 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift podName:243cb7f7-785c-4681-b344-ea4073562c84 nodeName:}" failed. No retries permitted until 2026-03-09 09:39:00.312224324 +0000 UTC m=+1141.136959905 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift") pod "swift-storage-0" (UID: "243cb7f7-785c-4681-b344-ea4073562c84") : configmap "swift-ring-files" not found Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.524498 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" event={"ID":"1eb95f27-3506-4519-8401-35420bdbb7e6","Type":"ContainerStarted","Data":"42d8ea28f99a69b068612b9a9e6a1cb73d12e022997a3c2c44f71846e299dddc"} Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.687211 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kv9nf"] Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.688860 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.717696 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gstn\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-kube-api-access-7gstn\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.718004 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.718034 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-run-httpd\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.718067 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-config-data\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.718381 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-log-httpd\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.721419 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kv9nf"] Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.819324 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gstn\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-kube-api-access-7gstn\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.819362 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.819391 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-run-httpd\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.819422 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-config-data\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.819467 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-log-httpd\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: E0309 09:38:58.819573 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:38:58 crc kubenswrapper[4692]: E0309 09:38:58.819607 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-kv9nf: configmap "swift-ring-files" not found Mar 09 09:38:58 crc kubenswrapper[4692]: E0309 09:38:58.819675 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift podName:1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a nodeName:}" failed. No retries permitted until 2026-03-09 09:38:59.319654877 +0000 UTC m=+1140.144390528 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift") pod "swift-proxy-76c998454c-kv9nf" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a") : configmap "swift-ring-files" not found Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.820026 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-run-httpd\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.820050 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-log-httpd\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.827256 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-config-data\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:58 crc kubenswrapper[4692]: I0309 09:38:58.844077 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gstn\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-kube-api-access-7gstn\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:59 crc kubenswrapper[4692]: I0309 09:38:59.337093 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:38:59 crc kubenswrapper[4692]: E0309 09:38:59.337319 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:38:59 crc kubenswrapper[4692]: E0309 09:38:59.337357 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-kv9nf: configmap "swift-ring-files" not found Mar 09 09:38:59 crc kubenswrapper[4692]: E0309 09:38:59.337418 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift podName:1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a nodeName:}" failed. No retries permitted until 2026-03-09 09:39:00.33740164 +0000 UTC m=+1141.162137221 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift") pod "swift-proxy-76c998454c-kv9nf" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a") : configmap "swift-ring-files" not found Mar 09 09:39:00 crc kubenswrapper[4692]: I0309 09:39:00.357936 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:00 crc kubenswrapper[4692]: I0309 09:39:00.358299 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:39:00 crc kubenswrapper[4692]: E0309 09:39:00.358144 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:39:00 crc kubenswrapper[4692]: E0309 09:39:00.358338 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-kv9nf: configmap "swift-ring-files" not found Mar 09 09:39:00 crc kubenswrapper[4692]: E0309 09:39:00.358407 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift podName:1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a nodeName:}" failed. No retries permitted until 2026-03-09 09:39:02.358382152 +0000 UTC m=+1143.183117733 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift") pod "swift-proxy-76c998454c-kv9nf" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a") : configmap "swift-ring-files" not found Mar 09 09:39:00 crc kubenswrapper[4692]: E0309 09:39:00.358527 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:39:00 crc kubenswrapper[4692]: E0309 09:39:00.358554 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:39:00 crc kubenswrapper[4692]: E0309 09:39:00.358609 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift podName:243cb7f7-785c-4681-b344-ea4073562c84 nodeName:}" failed. No retries permitted until 2026-03-09 09:39:04.358591908 +0000 UTC m=+1145.183327489 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift") pod "swift-storage-0" (UID: "243cb7f7-785c-4681-b344-ea4073562c84") : configmap "swift-ring-files" not found Mar 09 09:39:01 crc kubenswrapper[4692]: I0309 09:39:01.229303 4692 scope.go:117] "RemoveContainer" containerID="7322dbc30ea39903ba46c40fb451a821a0ca196163a623f7eaf62742fb37905d" Mar 09 09:39:02 crc kubenswrapper[4692]: I0309 09:39:02.387855 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:02 crc kubenswrapper[4692]: E0309 09:39:02.388054 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:39:02 crc kubenswrapper[4692]: E0309 09:39:02.388282 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-kv9nf: configmap "swift-ring-files" not found Mar 09 09:39:02 crc kubenswrapper[4692]: E0309 09:39:02.388354 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift podName:1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a nodeName:}" failed. No retries permitted until 2026-03-09 09:39:06.388330701 +0000 UTC m=+1147.213066332 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift") pod "swift-proxy-76c998454c-kv9nf" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a") : configmap "swift-ring-files" not found Mar 09 09:39:03 crc kubenswrapper[4692]: I0309 09:39:03.557042 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" event={"ID":"1eb95f27-3506-4519-8401-35420bdbb7e6","Type":"ContainerStarted","Data":"c20501be9e7f093efa0f3a1c0392cc53a297a64c0764c25306b40378d49d08e5"} Mar 09 09:39:03 crc kubenswrapper[4692]: I0309 09:39:03.584044 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" podStartSLOduration=2.870604485 podStartE2EDuration="7.584025995s" podCreationTimestamp="2026-03-09 09:38:56 +0000 UTC" firstStartedPulling="2026-03-09 09:38:57.866999702 +0000 UTC m=+1138.691735293" lastFinishedPulling="2026-03-09 09:39:02.580421222 +0000 UTC m=+1143.405156803" observedRunningTime="2026-03-09 09:39:03.579873473 +0000 UTC m=+1144.404609084" watchObservedRunningTime="2026-03-09 09:39:03.584025995 +0000 UTC m=+1144.408761576" Mar 09 09:39:04 crc kubenswrapper[4692]: I0309 09:39:04.423638 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:39:04 crc kubenswrapper[4692]: E0309 09:39:04.423941 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:39:04 crc kubenswrapper[4692]: E0309 09:39:04.424110 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:39:04 crc kubenswrapper[4692]: E0309 09:39:04.424232 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift podName:243cb7f7-785c-4681-b344-ea4073562c84 nodeName:}" failed. No retries permitted until 2026-03-09 09:39:12.42421032 +0000 UTC m=+1153.248945901 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift") pod "swift-storage-0" (UID: "243cb7f7-785c-4681-b344-ea4073562c84") : configmap "swift-ring-files" not found Mar 09 09:39:06 crc kubenswrapper[4692]: I0309 09:39:06.451534 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:06 crc kubenswrapper[4692]: E0309 09:39:06.451750 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:39:06 crc kubenswrapper[4692]: E0309 09:39:06.451764 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-proxy-76c998454c-kv9nf: configmap "swift-ring-files" not found Mar 09 09:39:06 crc kubenswrapper[4692]: E0309 09:39:06.451803 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift podName:1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a nodeName:}" failed. No retries permitted until 2026-03-09 09:39:14.451790527 +0000 UTC m=+1155.276526108 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift") pod "swift-proxy-76c998454c-kv9nf" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a") : configmap "swift-ring-files" not found Mar 09 09:39:10 crc kubenswrapper[4692]: I0309 09:39:10.615674 4692 generic.go:334] "Generic (PLEG): container finished" podID="1eb95f27-3506-4519-8401-35420bdbb7e6" containerID="c20501be9e7f093efa0f3a1c0392cc53a297a64c0764c25306b40378d49d08e5" exitCode=0 Mar 09 09:39:10 crc kubenswrapper[4692]: I0309 09:39:10.615751 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" event={"ID":"1eb95f27-3506-4519-8401-35420bdbb7e6","Type":"ContainerDied","Data":"c20501be9e7f093efa0f3a1c0392cc53a297a64c0764c25306b40378d49d08e5"} Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.899401 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.954412 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-dispersionconf\") pod \"1eb95f27-3506-4519-8401-35420bdbb7e6\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.954510 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-swiftconf\") pod \"1eb95f27-3506-4519-8401-35420bdbb7e6\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.954551 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eb95f27-3506-4519-8401-35420bdbb7e6-etc-swift\") pod \"1eb95f27-3506-4519-8401-35420bdbb7e6\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.954595 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9csm\" (UniqueName: \"kubernetes.io/projected/1eb95f27-3506-4519-8401-35420bdbb7e6-kube-api-access-b9csm\") pod \"1eb95f27-3506-4519-8401-35420bdbb7e6\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.954634 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-ring-data-devices\") pod \"1eb95f27-3506-4519-8401-35420bdbb7e6\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.954743 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-scripts\") pod \"1eb95f27-3506-4519-8401-35420bdbb7e6\" (UID: \"1eb95f27-3506-4519-8401-35420bdbb7e6\") " Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.955239 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1eb95f27-3506-4519-8401-35420bdbb7e6" (UID: "1eb95f27-3506-4519-8401-35420bdbb7e6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.955429 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1eb95f27-3506-4519-8401-35420bdbb7e6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1eb95f27-3506-4519-8401-35420bdbb7e6" (UID: "1eb95f27-3506-4519-8401-35420bdbb7e6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.960380 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb95f27-3506-4519-8401-35420bdbb7e6-kube-api-access-b9csm" (OuterVolumeSpecName: "kube-api-access-b9csm") pod "1eb95f27-3506-4519-8401-35420bdbb7e6" (UID: "1eb95f27-3506-4519-8401-35420bdbb7e6"). InnerVolumeSpecName "kube-api-access-b9csm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.962636 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1eb95f27-3506-4519-8401-35420bdbb7e6" (UID: "1eb95f27-3506-4519-8401-35420bdbb7e6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.973873 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-scripts" (OuterVolumeSpecName: "scripts") pod "1eb95f27-3506-4519-8401-35420bdbb7e6" (UID: "1eb95f27-3506-4519-8401-35420bdbb7e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:11 crc kubenswrapper[4692]: I0309 09:39:11.973903 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1eb95f27-3506-4519-8401-35420bdbb7e6" (UID: "1eb95f27-3506-4519-8401-35420bdbb7e6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.056780 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.056819 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.056835 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1eb95f27-3506-4519-8401-35420bdbb7e6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.056846 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1eb95f27-3506-4519-8401-35420bdbb7e6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.056857 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9csm\" (UniqueName: \"kubernetes.io/projected/1eb95f27-3506-4519-8401-35420bdbb7e6-kube-api-access-b9csm\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.056870 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1eb95f27-3506-4519-8401-35420bdbb7e6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.463914 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.468333 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"swift-storage-0\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.632232 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" event={"ID":"1eb95f27-3506-4519-8401-35420bdbb7e6","Type":"ContainerDied","Data":"42d8ea28f99a69b068612b9a9e6a1cb73d12e022997a3c2c44f71846e299dddc"} Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.632276 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-z69jg" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.632276 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42d8ea28f99a69b068612b9a9e6a1cb73d12e022997a3c2c44f71846e299dddc" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.734337 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:39:12 crc kubenswrapper[4692]: I0309 09:39:12.825640 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:13 crc kubenswrapper[4692]: I0309 09:39:13.226968 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:39:13 crc kubenswrapper[4692]: I0309 09:39:13.638656 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"9ff78cf169b13930889ded5fc90d081b4e3777328901c64edbbe2681eedc2153"} Mar 09 09:39:14 crc kubenswrapper[4692]: I0309 09:39:14.453760 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:14 crc kubenswrapper[4692]: I0309 09:39:14.490396 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:14 crc kubenswrapper[4692]: I0309 09:39:14.497202 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"swift-proxy-76c998454c-kv9nf\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:14 crc kubenswrapper[4692]: I0309 09:39:14.604634 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:14 crc kubenswrapper[4692]: I0309 09:39:14.672914 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"d9fca122f4493bee53c9b8f344da247b633bac2de061a56c8724e98fc1516460"} Mar 09 09:39:14 crc kubenswrapper[4692]: I0309 09:39:14.672956 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"cbf6beff918423c3958f88e8749e875ff0a036cc342ad42299854971e4745a5a"} Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.205084 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kv9nf"] Mar 09 09:39:15 crc kubenswrapper[4692]: W0309 09:39:15.216324 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e99c1fd_25a0_4f63_9eab_b9dddcbfc39a.slice/crio-4efb13ce1cc00883326f4aad8e1e720264d37cbe0a5d6006f1a9cf67646b1166 WatchSource:0}: Error finding container 4efb13ce1cc00883326f4aad8e1e720264d37cbe0a5d6006f1a9cf67646b1166: Status 404 returned error can't find the container with id 4efb13ce1cc00883326f4aad8e1e720264d37cbe0a5d6006f1a9cf67646b1166 Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.682732 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"704ea831492a29f5a461f465943223a5cf5a1275ba375cc94f8c0af1c8c9c7dc"} Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.682773 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"ccb5034272c4443a5d30c3e77c9d8499de513d71548f5dfc4aa471c4cf783678"} Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.685014 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" event={"ID":"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a","Type":"ContainerStarted","Data":"099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1"} Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.685056 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" event={"ID":"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a","Type":"ContainerStarted","Data":"d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4"} Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.685067 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" event={"ID":"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a","Type":"ContainerStarted","Data":"4efb13ce1cc00883326f4aad8e1e720264d37cbe0a5d6006f1a9cf67646b1166"} Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.685207 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:15 crc kubenswrapper[4692]: I0309 09:39:15.708709 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" podStartSLOduration=17.708694136 podStartE2EDuration="17.708694136s" podCreationTimestamp="2026-03-09 09:38:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:39:15.706013934 +0000 UTC m=+1156.530749525" watchObservedRunningTime="2026-03-09 09:39:15.708694136 +0000 UTC m=+1156.533429717" Mar 09 09:39:16 crc kubenswrapper[4692]: I0309 09:39:16.032456 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:16 crc kubenswrapper[4692]: I0309 09:39:16.695449 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"dfc762ad83fc57e117251ed04fa58e82ee932457ae195cc07a45d8582c6f48b3"} Mar 09 09:39:16 crc kubenswrapper[4692]: I0309 09:39:16.695531 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"7d475f88fe15d82cbca79a9f0f44b9547c75bd6a08ec54a080ad68c5c9eb1667"} Mar 09 09:39:16 crc kubenswrapper[4692]: I0309 09:39:16.695550 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:17 crc kubenswrapper[4692]: I0309 09:39:17.624216 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:39:17 crc kubenswrapper[4692]: I0309 09:39:17.624595 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:39:17 crc kubenswrapper[4692]: I0309 09:39:17.733795 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"0574622c3ea59457671ae2d3e5912c7907129a0385b9617eb6c635283f6aedf8"} Mar 09 09:39:17 crc kubenswrapper[4692]: I0309 09:39:17.733833 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"4dcd159606ea3eb0a56b35c2bbc8955b8cf4ee1aeaddc4a57f92fb1b5c579662"} Mar 09 09:39:17 crc kubenswrapper[4692]: I0309 09:39:17.742482 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:19 crc kubenswrapper[4692]: I0309 09:39:19.367052 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:19 crc kubenswrapper[4692]: I0309 09:39:19.831459 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"099b4f3feffdd401f91b122c7eebf222dc20cb7003286e6b3fdbb5b1dd988b2b"} Mar 09 09:39:19 crc kubenswrapper[4692]: I0309 09:39:19.831509 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"171d835038bf3f61e38646d3ac1a693d257467e87de91754cb63ad379961e59e"} Mar 09 09:39:19 crc kubenswrapper[4692]: I0309 09:39:19.831523 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"26c226cab26d728919ade2a94001899645db4128cc63122ef0b7e131dafe18e4"} Mar 09 09:39:20 crc kubenswrapper[4692]: I0309 09:39:20.847905 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"86718cd93961aca462a530e4a4791cb7e77a447add0e2a72803c9f229c6ede23"} Mar 09 09:39:20 crc kubenswrapper[4692]: I0309 09:39:20.848218 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"140a3b0ef60391e40a7e1b3dbe1d9e95bbabab2391fdaa782882ae34c7dade4a"} Mar 09 09:39:20 crc kubenswrapper[4692]: I0309 09:39:20.848232 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"5daee84511b7c0cbba039fd3cbcf9d60634e7a175565d403aa95d2d6bc919b4c"} Mar 09 09:39:20 crc kubenswrapper[4692]: I0309 09:39:20.848243 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerStarted","Data":"b3f6440db1f20d77b29490fabaa341162a2757de5636fde37edaedeb487ba9e8"} Mar 09 09:39:20 crc kubenswrapper[4692]: I0309 09:39:20.883810 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-0" podStartSLOduration=19.838631804 podStartE2EDuration="25.883788495s" podCreationTimestamp="2026-03-09 09:38:55 +0000 UTC" firstStartedPulling="2026-03-09 09:39:13.235608141 +0000 UTC m=+1154.060343722" lastFinishedPulling="2026-03-09 09:39:19.280764832 +0000 UTC m=+1160.105500413" observedRunningTime="2026-03-09 09:39:20.879886039 +0000 UTC m=+1161.704621630" watchObservedRunningTime="2026-03-09 09:39:20.883788495 +0000 UTC m=+1161.708524076" Mar 09 09:39:20 crc kubenswrapper[4692]: I0309 09:39:20.942150 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:22 crc kubenswrapper[4692]: I0309 09:39:22.450804 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:23 crc kubenswrapper[4692]: I0309 09:39:23.982990 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:24 crc kubenswrapper[4692]: I0309 09:39:24.608700 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:24 crc kubenswrapper[4692]: I0309 09:39:24.610958 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:25 crc kubenswrapper[4692]: I0309 09:39:25.526046 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:27 crc kubenswrapper[4692]: I0309 09:39:27.034301 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-z69jg_1eb95f27-3506-4519-8401-35420bdbb7e6/swift-ring-rebalance/0.log" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.400326 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:39:28 crc kubenswrapper[4692]: E0309 09:39:28.400672 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb95f27-3506-4519-8401-35420bdbb7e6" containerName="swift-ring-rebalance" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.400684 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb95f27-3506-4519-8401-35420bdbb7e6" containerName="swift-ring-rebalance" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.400822 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb95f27-3506-4519-8401-35420bdbb7e6" containerName="swift-ring-rebalance" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.404657 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.404945 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.410966 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.418208 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.426107 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451464 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-cache\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451516 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-etc-swift\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451540 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-etc-swift\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451561 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451613 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-lock\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451632 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-cache\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451651 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451679 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrfvn\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-kube-api-access-zrfvn\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451704 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-lock\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.451748 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z925c\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-kube-api-access-z925c\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553267 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-lock\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553381 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z925c\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-kube-api-access-z925c\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553474 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-cache\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553528 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-etc-swift\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553550 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-etc-swift\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553634 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553734 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-lock\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553786 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-cache\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553809 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.553863 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrfvn\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-kube-api-access-zrfvn\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.554475 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") device mount path \"/mnt/openstack/pv07\"" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.554499 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") device mount path \"/mnt/openstack/pv09\"" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.554768 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-lock\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.554937 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-cache\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.555231 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-lock\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.555566 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-cache\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.566733 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-etc-swift\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.566875 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-etc-swift\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.576049 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z925c\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-kube-api-access-z925c\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.576242 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-2\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.580390 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrfvn\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-kube-api-access-zrfvn\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.587902 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-1\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.741328 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.758627 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.769184 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-z69jg"] Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.776454 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-z69jg"] Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.795311 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-cl5cc"] Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.796355 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.801332 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.801454 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-cl5cc"] Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.801514 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.961272 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-scripts\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.961642 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-dispersionconf\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.961664 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-ring-data-devices\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.961697 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn65d\" (UniqueName: \"kubernetes.io/projected/23a8c09b-c47b-4834-972d-437aec64bbc8-kube-api-access-cn65d\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.961722 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-swiftconf\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:28 crc kubenswrapper[4692]: I0309 09:39:28.961776 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/23a8c09b-c47b-4834-972d-437aec64bbc8-etc-swift\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.064278 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-scripts\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.064399 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-dispersionconf\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.064445 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-ring-data-devices\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.064492 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn65d\" (UniqueName: \"kubernetes.io/projected/23a8c09b-c47b-4834-972d-437aec64bbc8-kube-api-access-cn65d\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.064524 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-swiftconf\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.064552 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/23a8c09b-c47b-4834-972d-437aec64bbc8-etc-swift\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.068591 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-scripts\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.070028 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-ring-data-devices\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.070449 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/23a8c09b-c47b-4834-972d-437aec64bbc8-etc-swift\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.072262 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-swiftconf\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.072680 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-dispersionconf\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.092465 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn65d\" (UniqueName: \"kubernetes.io/projected/23a8c09b-c47b-4834-972d-437aec64bbc8-kube-api-access-cn65d\") pod \"swift-ring-rebalance-cl5cc\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.145870 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.146479 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:39:29 crc kubenswrapper[4692]: W0309 09:39:29.160883 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6904b3a2_2562_46b8_8c7c_634d9c7f917e.slice/crio-81269b346e9df9f2f6656d2e4deab5af9b59db56c995ea774262cb484f0bb208 WatchSource:0}: Error finding container 81269b346e9df9f2f6656d2e4deab5af9b59db56c995ea774262cb484f0bb208: Status 404 returned error can't find the container with id 81269b346e9df9f2f6656d2e4deab5af9b59db56c995ea774262cb484f0bb208 Mar 09 09:39:29 crc kubenswrapper[4692]: I0309 09:39:29.297752 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:29.631145 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-cl5cc"] Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.028616 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4"} Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.028933 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796"} Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.028949 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"c8ae29f7cb9bb78b1d1be941154615915a0a3b76626caa822de4d652bb6226a0"} Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.031054 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" event={"ID":"23a8c09b-c47b-4834-972d-437aec64bbc8","Type":"ContainerStarted","Data":"39edcc2846edb830546f11a10dcf38d264cd68f0c2314d0e12db14c27302b639"} Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.043607 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"a397709beb06ddc2e3731ee791524a484e7b3ac4a47a134ded55ece54ff1c7b7"} Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.043651 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"96a774aea1b785c85e9e4db9fdf687a4a017ec236987833423a664e390b9a635"} Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.043666 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"81269b346e9df9f2f6656d2e4deab5af9b59db56c995ea774262cb484f0bb208"} Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.127798 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb95f27-3506-4519-8401-35420bdbb7e6" path="/var/lib/kubelet/pods/1eb95f27-3506-4519-8401-35420bdbb7e6/volumes" Mar 09 09:39:30 crc kubenswrapper[4692]: I0309 09:39:30.202650 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" podStartSLOduration=2.202628374 podStartE2EDuration="2.202628374s" podCreationTimestamp="2026-03-09 09:39:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:39:30.062152781 +0000 UTC m=+1170.886888362" watchObservedRunningTime="2026-03-09 09:39:30.202628374 +0000 UTC m=+1171.027363955" Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.057725 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.057881 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.057891 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.057900 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.058939 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" event={"ID":"23a8c09b-c47b-4834-972d-437aec64bbc8","Type":"ContainerStarted","Data":"8eabfccd02936e226e1669133e8352835a91c3b973895cb307ea66b9d7fa12d1"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.064017 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"caa19051a64d2139ffcaae693702a9add912d133e2c6ad1aa105da66652c3dbf"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.064059 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"dcb1d92e14edd56619ba2577f2bd3ce1ada590038150f3acc568b1fa13bedae4"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.064070 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"7f36f35de0e6d3a1b838b67e886b8a52e9398ce82c050e5b935cf0379ad62d86"} Mar 09 09:39:31 crc kubenswrapper[4692]: I0309 09:39:31.064087 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"6ab00368cc89fcf8161f0d43ba1d20c067115c63db8f87982937f03a141ce1a2"} Mar 09 09:39:32 crc kubenswrapper[4692]: I0309 09:39:32.193139 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"38075b27427df3883c9982c59fdeffdb66ca032b3966023e619377f2e77b8c9d"} Mar 09 09:39:32 crc kubenswrapper[4692]: I0309 09:39:32.193446 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"966a843ea31860db24f4b9279d146c5d253cc277653ba196f6ef2a6887360cb4"} Mar 09 09:39:32 crc kubenswrapper[4692]: I0309 09:39:32.266974 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c"} Mar 09 09:39:32 crc kubenswrapper[4692]: I0309 09:39:32.267013 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d"} Mar 09 09:39:32 crc kubenswrapper[4692]: I0309 09:39:32.267025 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.296863 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"0321f9472e248b304e48ec950df8cd09d170ae9fdc38f48e874938cd6e895e85"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.296902 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"46cbedf0abfd3f31c376d3f0fc737f9687de69d123ae655da80d749a3fc71176"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.296916 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"13f1d2241b7ac6072e93443f2565800d1acf3174a4de209a3cdd382ce7b2c893"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.296927 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"f6a023b7f054ee3eb28ad8d00763fbf33ab4b36409454d63e42ea87e4a6e617f"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.302600 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.302652 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.302665 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016"} Mar 09 09:39:33 crc kubenswrapper[4692]: I0309 09:39:33.302675 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c"} Mar 09 09:39:34 crc kubenswrapper[4692]: I0309 09:39:34.360917 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba"} Mar 09 09:39:34 crc kubenswrapper[4692]: I0309 09:39:34.364320 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"06646bb3d2ff65eedaafa404d7d0aa7923592978978ae2271af7f090c46dc8fe"} Mar 09 09:39:35 crc kubenswrapper[4692]: I0309 09:39:35.379497 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"280979651d624f8b196a49bb676333d78dd6e3ba38a2c3044d7159baa49a045b"} Mar 09 09:39:35 crc kubenswrapper[4692]: I0309 09:39:35.379548 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerStarted","Data":"328c19774f97092a3e3f498eec5b95de21679577a75dd683d879de5fa9cefbe6"} Mar 09 09:39:35 crc kubenswrapper[4692]: I0309 09:39:35.387259 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerStarted","Data":"75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d"} Mar 09 09:39:35 crc kubenswrapper[4692]: I0309 09:39:35.433502 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-1" podStartSLOduration=8.433479222 podStartE2EDuration="8.433479222s" podCreationTimestamp="2026-03-09 09:39:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:39:35.428214369 +0000 UTC m=+1176.252949980" watchObservedRunningTime="2026-03-09 09:39:35.433479222 +0000 UTC m=+1176.258214803" Mar 09 09:39:35 crc kubenswrapper[4692]: I0309 09:39:35.471075 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-2" podStartSLOduration=8.471054776999999 podStartE2EDuration="8.471054777s" podCreationTimestamp="2026-03-09 09:39:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:39:35.465742724 +0000 UTC m=+1176.290478335" watchObservedRunningTime="2026-03-09 09:39:35.471054777 +0000 UTC m=+1176.295790358" Mar 09 09:39:42 crc kubenswrapper[4692]: I0309 09:39:42.439677 4692 generic.go:334] "Generic (PLEG): container finished" podID="23a8c09b-c47b-4834-972d-437aec64bbc8" containerID="8eabfccd02936e226e1669133e8352835a91c3b973895cb307ea66b9d7fa12d1" exitCode=0 Mar 09 09:39:42 crc kubenswrapper[4692]: I0309 09:39:42.439762 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" event={"ID":"23a8c09b-c47b-4834-972d-437aec64bbc8","Type":"ContainerDied","Data":"8eabfccd02936e226e1669133e8352835a91c3b973895cb307ea66b9d7fa12d1"} Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.804291 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.972195 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-dispersionconf\") pod \"23a8c09b-c47b-4834-972d-437aec64bbc8\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.972632 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-scripts\") pod \"23a8c09b-c47b-4834-972d-437aec64bbc8\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.972701 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-ring-data-devices\") pod \"23a8c09b-c47b-4834-972d-437aec64bbc8\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.972731 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/23a8c09b-c47b-4834-972d-437aec64bbc8-etc-swift\") pod \"23a8c09b-c47b-4834-972d-437aec64bbc8\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.972787 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn65d\" (UniqueName: \"kubernetes.io/projected/23a8c09b-c47b-4834-972d-437aec64bbc8-kube-api-access-cn65d\") pod \"23a8c09b-c47b-4834-972d-437aec64bbc8\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.972884 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-swiftconf\") pod \"23a8c09b-c47b-4834-972d-437aec64bbc8\" (UID: \"23a8c09b-c47b-4834-972d-437aec64bbc8\") " Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.973524 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "23a8c09b-c47b-4834-972d-437aec64bbc8" (UID: "23a8c09b-c47b-4834-972d-437aec64bbc8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.973717 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23a8c09b-c47b-4834-972d-437aec64bbc8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "23a8c09b-c47b-4834-972d-437aec64bbc8" (UID: "23a8c09b-c47b-4834-972d-437aec64bbc8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.977614 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23a8c09b-c47b-4834-972d-437aec64bbc8-kube-api-access-cn65d" (OuterVolumeSpecName: "kube-api-access-cn65d") pod "23a8c09b-c47b-4834-972d-437aec64bbc8" (UID: "23a8c09b-c47b-4834-972d-437aec64bbc8"). InnerVolumeSpecName "kube-api-access-cn65d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.996915 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-scripts" (OuterVolumeSpecName: "scripts") pod "23a8c09b-c47b-4834-972d-437aec64bbc8" (UID: "23a8c09b-c47b-4834-972d-437aec64bbc8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:43 crc kubenswrapper[4692]: I0309 09:39:43.998994 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "23a8c09b-c47b-4834-972d-437aec64bbc8" (UID: "23a8c09b-c47b-4834-972d-437aec64bbc8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.004761 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "23a8c09b-c47b-4834-972d-437aec64bbc8" (UID: "23a8c09b-c47b-4834-972d-437aec64bbc8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.074987 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn65d\" (UniqueName: \"kubernetes.io/projected/23a8c09b-c47b-4834-972d-437aec64bbc8-kube-api-access-cn65d\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.075023 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.075035 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/23a8c09b-c47b-4834-972d-437aec64bbc8-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.075046 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.075057 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/23a8c09b-c47b-4834-972d-437aec64bbc8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.075104 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/23a8c09b-c47b-4834-972d-437aec64bbc8-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.461403 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.461318 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-cl5cc" event={"ID":"23a8c09b-c47b-4834-972d-437aec64bbc8","Type":"ContainerDied","Data":"39edcc2846edb830546f11a10dcf38d264cd68f0c2314d0e12db14c27302b639"} Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.461459 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39edcc2846edb830546f11a10dcf38d264cd68f0c2314d0e12db14c27302b639" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.683446 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7"] Mar 09 09:39:44 crc kubenswrapper[4692]: E0309 09:39:44.683854 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23a8c09b-c47b-4834-972d-437aec64bbc8" containerName="swift-ring-rebalance" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.683883 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="23a8c09b-c47b-4834-972d-437aec64bbc8" containerName="swift-ring-rebalance" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.684091 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="23a8c09b-c47b-4834-972d-437aec64bbc8" containerName="swift-ring-rebalance" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.684732 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.688277 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.688397 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.693315 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7"] Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.786135 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlhg9\" (UniqueName: \"kubernetes.io/projected/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-kube-api-access-jlhg9\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.786203 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-scripts\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.786228 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-etc-swift\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.786265 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-ring-data-devices\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.786285 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-swiftconf\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.786362 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-dispersionconf\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.888322 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlhg9\" (UniqueName: \"kubernetes.io/projected/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-kube-api-access-jlhg9\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.888393 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-scripts\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.888431 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-etc-swift\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.888470 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-ring-data-devices\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.888503 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-swiftconf\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.888576 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-dispersionconf\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.889940 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-etc-swift\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.889973 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-scripts\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.890232 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-ring-data-devices\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.894854 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-swiftconf\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.898595 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-dispersionconf\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:44 crc kubenswrapper[4692]: I0309 09:39:44.904059 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlhg9\" (UniqueName: \"kubernetes.io/projected/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-kube-api-access-jlhg9\") pod \"swift-ring-rebalance-debug-k7gs7\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:45 crc kubenswrapper[4692]: I0309 09:39:45.003210 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:45 crc kubenswrapper[4692]: I0309 09:39:45.412976 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7"] Mar 09 09:39:45 crc kubenswrapper[4692]: W0309 09:39:45.416795 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab8c298b_e372_401d_a6f3_9ad8f92e24f3.slice/crio-fe72415a39a64959c55de18cc9561f1e657773dc07df2aadfc00f58c49ce3543 WatchSource:0}: Error finding container fe72415a39a64959c55de18cc9561f1e657773dc07df2aadfc00f58c49ce3543: Status 404 returned error can't find the container with id fe72415a39a64959c55de18cc9561f1e657773dc07df2aadfc00f58c49ce3543 Mar 09 09:39:45 crc kubenswrapper[4692]: I0309 09:39:45.470354 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" event={"ID":"ab8c298b-e372-401d-a6f3-9ad8f92e24f3","Type":"ContainerStarted","Data":"fe72415a39a64959c55de18cc9561f1e657773dc07df2aadfc00f58c49ce3543"} Mar 09 09:39:46 crc kubenswrapper[4692]: I0309 09:39:46.482755 4692 generic.go:334] "Generic (PLEG): container finished" podID="ab8c298b-e372-401d-a6f3-9ad8f92e24f3" containerID="06647ab5571a3f1c36fdaeb677d5e716f3e9e42456f0aa75873178becc724a2c" exitCode=0 Mar 09 09:39:46 crc kubenswrapper[4692]: I0309 09:39:46.483087 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" event={"ID":"ab8c298b-e372-401d-a6f3-9ad8f92e24f3","Type":"ContainerDied","Data":"06647ab5571a3f1c36fdaeb677d5e716f3e9e42456f0aa75873178becc724a2c"} Mar 09 09:39:46 crc kubenswrapper[4692]: I0309 09:39:46.526392 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7"] Mar 09 09:39:46 crc kubenswrapper[4692]: I0309 09:39:46.531105 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7"] Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.624346 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.624419 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.624471 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.625191 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a00e663af9bd3bc83fcf84afc788c06c37b9025a209fe4f4ecbe66f9658cebf6"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.625266 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://a00e663af9bd3bc83fcf84afc788c06c37b9025a209fe4f4ecbe66f9658cebf6" gracePeriod=600 Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.784449 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.895351 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg"] Mar 09 09:39:47 crc kubenswrapper[4692]: E0309 09:39:47.895738 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab8c298b-e372-401d-a6f3-9ad8f92e24f3" containerName="swift-ring-rebalance" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.895763 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab8c298b-e372-401d-a6f3-9ad8f92e24f3" containerName="swift-ring-rebalance" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.895973 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab8c298b-e372-401d-a6f3-9ad8f92e24f3" containerName="swift-ring-rebalance" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.896617 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.903499 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg"] Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.929723 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-etc-swift\") pod \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.929801 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-scripts\") pod \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.929831 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-dispersionconf\") pod \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.929876 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-ring-data-devices\") pod \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.929948 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlhg9\" (UniqueName: \"kubernetes.io/projected/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-kube-api-access-jlhg9\") pod \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.930016 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-swiftconf\") pod \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\" (UID: \"ab8c298b-e372-401d-a6f3-9ad8f92e24f3\") " Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.931088 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ab8c298b-e372-401d-a6f3-9ad8f92e24f3" (UID: "ab8c298b-e372-401d-a6f3-9ad8f92e24f3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.931081 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ab8c298b-e372-401d-a6f3-9ad8f92e24f3" (UID: "ab8c298b-e372-401d-a6f3-9ad8f92e24f3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.936624 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-kube-api-access-jlhg9" (OuterVolumeSpecName: "kube-api-access-jlhg9") pod "ab8c298b-e372-401d-a6f3-9ad8f92e24f3" (UID: "ab8c298b-e372-401d-a6f3-9ad8f92e24f3"). InnerVolumeSpecName "kube-api-access-jlhg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.951647 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-scripts" (OuterVolumeSpecName: "scripts") pod "ab8c298b-e372-401d-a6f3-9ad8f92e24f3" (UID: "ab8c298b-e372-401d-a6f3-9ad8f92e24f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.956059 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ab8c298b-e372-401d-a6f3-9ad8f92e24f3" (UID: "ab8c298b-e372-401d-a6f3-9ad8f92e24f3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:47 crc kubenswrapper[4692]: I0309 09:39:47.966808 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ab8c298b-e372-401d-a6f3-9ad8f92e24f3" (UID: "ab8c298b-e372-401d-a6f3-9ad8f92e24f3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.031581 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-dispersionconf\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.031937 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79ba58b3-a4a4-4375-b1e2-285f952e08a4-etc-swift\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.031999 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-ring-data-devices\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032028 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-scripts\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032109 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-swiftconf\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032259 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdqxz\" (UniqueName: \"kubernetes.io/projected/79ba58b3-a4a4-4375-b1e2-285f952e08a4-kube-api-access-zdqxz\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032485 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032508 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032522 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032536 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032547 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.032557 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlhg9\" (UniqueName: \"kubernetes.io/projected/ab8c298b-e372-401d-a6f3-9ad8f92e24f3-kube-api-access-jlhg9\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.081450 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab8c298b-e372-401d-a6f3-9ad8f92e24f3" path="/var/lib/kubelet/pods/ab8c298b-e372-401d-a6f3-9ad8f92e24f3/volumes" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.134618 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-ring-data-devices\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.134671 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-scripts\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.134727 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-swiftconf\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.134764 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdqxz\" (UniqueName: \"kubernetes.io/projected/79ba58b3-a4a4-4375-b1e2-285f952e08a4-kube-api-access-zdqxz\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.134805 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-dispersionconf\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.135012 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79ba58b3-a4a4-4375-b1e2-285f952e08a4-etc-swift\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.135545 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-scripts\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.135571 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79ba58b3-a4a4-4375-b1e2-285f952e08a4-etc-swift\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.135998 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-ring-data-devices\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.138654 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-dispersionconf\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.138819 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-swiftconf\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.154853 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdqxz\" (UniqueName: \"kubernetes.io/projected/79ba58b3-a4a4-4375-b1e2-285f952e08a4-kube-api-access-zdqxz\") pod \"swift-ring-rebalance-debug-n4sfg\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.217970 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.470276 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg"] Mar 09 09:39:48 crc kubenswrapper[4692]: W0309 09:39:48.481003 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79ba58b3_a4a4_4375_b1e2_285f952e08a4.slice/crio-7266bf7e3e131a786a6202b04b279400494e398552ad63527f51bad67ae7f736 WatchSource:0}: Error finding container 7266bf7e3e131a786a6202b04b279400494e398552ad63527f51bad67ae7f736: Status 404 returned error can't find the container with id 7266bf7e3e131a786a6202b04b279400494e398552ad63527f51bad67ae7f736 Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.499294 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="a00e663af9bd3bc83fcf84afc788c06c37b9025a209fe4f4ecbe66f9658cebf6" exitCode=0 Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.499362 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"a00e663af9bd3bc83fcf84afc788c06c37b9025a209fe4f4ecbe66f9658cebf6"} Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.499406 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"dc5528f2b79383c92cd6d5389f75e574b39b690b0caddba49d4c6810fb3c85bb"} Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.499454 4692 scope.go:117] "RemoveContainer" containerID="c7bb4acc73ba36420c0329ef9cb241fb590012d44ec365a56134ad986b42e9b2" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.502987 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-k7gs7" Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.507965 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" event={"ID":"79ba58b3-a4a4-4375-b1e2-285f952e08a4","Type":"ContainerStarted","Data":"7266bf7e3e131a786a6202b04b279400494e398552ad63527f51bad67ae7f736"} Mar 09 09:39:48 crc kubenswrapper[4692]: I0309 09:39:48.535122 4692 scope.go:117] "RemoveContainer" containerID="06647ab5571a3f1c36fdaeb677d5e716f3e9e42456f0aa75873178becc724a2c" Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.515714 4692 generic.go:334] "Generic (PLEG): container finished" podID="79ba58b3-a4a4-4375-b1e2-285f952e08a4" containerID="b38c6942b9a7bf311d47525695855d615ccb8e90ebcb397ff550fd91174281bc" exitCode=0 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.515901 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" event={"ID":"79ba58b3-a4a4-4375-b1e2-285f952e08a4","Type":"ContainerDied","Data":"b38c6942b9a7bf311d47525695855d615ccb8e90ebcb397ff550fd91174281bc"} Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.549735 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.555806 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.673014 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-cl5cc"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.682011 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.683462 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="swift-recon-cron" containerID="cri-o://280979651d624f8b196a49bb676333d78dd6e3ba38a2c3044d7159baa49a045b" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.683980 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-server" containerID="cri-o://96a774aea1b785c85e9e4db9fdf687a4a017ec236987833423a664e390b9a635" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684031 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-updater" containerID="cri-o://38075b27427df3883c9982c59fdeffdb66ca032b3966023e619377f2e77b8c9d" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684093 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="rsync" containerID="cri-o://328c19774f97092a3e3f498eec5b95de21679577a75dd683d879de5fa9cefbe6" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684142 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-replicator" containerID="cri-o://13f1d2241b7ac6072e93443f2565800d1acf3174a4de209a3cdd382ce7b2c893" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684131 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-auditor" containerID="cri-o://966a843ea31860db24f4b9279d146c5d253cc277653ba196f6ef2a6887360cb4" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684192 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-updater" containerID="cri-o://0321f9472e248b304e48ec950df8cd09d170ae9fdc38f48e874938cd6e895e85" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684211 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-server" containerID="cri-o://f6a023b7f054ee3eb28ad8d00763fbf33ab4b36409454d63e42ea87e4a6e617f" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684232 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-reaper" containerID="cri-o://6ab00368cc89fcf8161f0d43ba1d20c067115c63db8f87982937f03a141ce1a2" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684214 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-auditor" containerID="cri-o://caa19051a64d2139ffcaae693702a9add912d133e2c6ad1aa105da66652c3dbf" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684259 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-replicator" containerID="cri-o://a397709beb06ddc2e3731ee791524a484e7b3ac4a47a134ded55ece54ff1c7b7" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684149 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-auditor" containerID="cri-o://46cbedf0abfd3f31c376d3f0fc737f9687de69d123ae655da80d749a3fc71176" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684183 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-replicator" containerID="cri-o://dcb1d92e14edd56619ba2577f2bd3ce1ada590038150f3acc568b1fa13bedae4" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684208 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-expirer" containerID="cri-o://06646bb3d2ff65eedaafa404d7d0aa7923592978978ae2271af7f090c46dc8fe" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.684100 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-1" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-server" containerID="cri-o://7f36f35de0e6d3a1b838b67e886b8a52e9398ce82c050e5b935cf0379ad62d86" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.699721 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700257 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-server" containerID="cri-o://cbf6beff918423c3958f88e8749e875ff0a036cc342ad42299854971e4745a5a" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700675 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="swift-recon-cron" containerID="cri-o://86718cd93961aca462a530e4a4791cb7e77a447add0e2a72803c9f229c6ede23" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700738 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="rsync" containerID="cri-o://140a3b0ef60391e40a7e1b3dbe1d9e95bbabab2391fdaa782882ae34c7dade4a" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700788 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-expirer" containerID="cri-o://5daee84511b7c0cbba039fd3cbcf9d60634e7a175565d403aa95d2d6bc919b4c" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700833 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-updater" containerID="cri-o://b3f6440db1f20d77b29490fabaa341162a2757de5636fde37edaedeb487ba9e8" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700881 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-auditor" containerID="cri-o://099b4f3feffdd401f91b122c7eebf222dc20cb7003286e6b3fdbb5b1dd988b2b" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700926 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-replicator" containerID="cri-o://171d835038bf3f61e38646d3ac1a693d257467e87de91754cb63ad379961e59e" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.700966 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-server" containerID="cri-o://26c226cab26d728919ade2a94001899645db4128cc63122ef0b7e131dafe18e4" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.701023 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-updater" containerID="cri-o://0574622c3ea59457671ae2d3e5912c7907129a0385b9617eb6c635283f6aedf8" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.701070 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-auditor" containerID="cri-o://4dcd159606ea3eb0a56b35c2bbc8955b8cf4ee1aeaddc4a57f92fb1b5c579662" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.701113 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-replicator" containerID="cri-o://dfc762ad83fc57e117251ed04fa58e82ee932457ae195cc07a45d8582c6f48b3" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.701186 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-server" containerID="cri-o://7d475f88fe15d82cbca79a9f0f44b9547c75bd6a08ec54a080ad68c5c9eb1667" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.701236 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-reaper" containerID="cri-o://704ea831492a29f5a461f465943223a5cf5a1275ba375cc94f8c0af1c8c9c7dc" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.701284 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-auditor" containerID="cri-o://ccb5034272c4443a5d30c3e77c9d8499de513d71548f5dfc4aa471c4cf783678" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.701328 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-replicator" containerID="cri-o://d9fca122f4493bee53c9b8f344da247b633bac2de061a56c8724e98fc1516460" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.771178 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.772927 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="swift-recon-cron" containerID="cri-o://75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.772717 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-server" containerID="cri-o://8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773117 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-replicator" containerID="cri-o://68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773093 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="rsync" containerID="cri-o://2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773253 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-server" containerID="cri-o://53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773271 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-updater" containerID="cri-o://dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773240 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-expirer" containerID="cri-o://ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773373 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-reaper" containerID="cri-o://ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773375 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-auditor" containerID="cri-o://5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773187 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-auditor" containerID="cri-o://24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773449 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-auditor" containerID="cri-o://8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773466 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-replicator" containerID="cri-o://c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773543 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-updater" containerID="cri-o://a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773522 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-server" containerID="cri-o://1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.773649 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-2" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-replicator" containerID="cri-o://462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.846823 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-cl5cc"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.858377 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kv9nf"] Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.858727 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-httpd" containerID="cri-o://d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4" gracePeriod=30 Mar 09 09:39:49 crc kubenswrapper[4692]: I0309 09:39:49.859296 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-server" containerID="cri-o://099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1" gracePeriod=30 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.119888 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23a8c09b-c47b-4834-972d-437aec64bbc8" path="/var/lib/kubelet/pods/23a8c09b-c47b-4834-972d-437aec64bbc8/volumes" Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538055 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="328c19774f97092a3e3f498eec5b95de21679577a75dd683d879de5fa9cefbe6" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538377 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="06646bb3d2ff65eedaafa404d7d0aa7923592978978ae2271af7f090c46dc8fe" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538387 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="0321f9472e248b304e48ec950df8cd09d170ae9fdc38f48e874938cd6e895e85" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538396 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="46cbedf0abfd3f31c376d3f0fc737f9687de69d123ae655da80d749a3fc71176" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538403 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="13f1d2241b7ac6072e93443f2565800d1acf3174a4de209a3cdd382ce7b2c893" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538410 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="f6a023b7f054ee3eb28ad8d00763fbf33ab4b36409454d63e42ea87e4a6e617f" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538417 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="38075b27427df3883c9982c59fdeffdb66ca032b3966023e619377f2e77b8c9d" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538424 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="966a843ea31860db24f4b9279d146c5d253cc277653ba196f6ef2a6887360cb4" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538432 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="dcb1d92e14edd56619ba2577f2bd3ce1ada590038150f3acc568b1fa13bedae4" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538439 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="7f36f35de0e6d3a1b838b67e886b8a52e9398ce82c050e5b935cf0379ad62d86" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538448 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="6ab00368cc89fcf8161f0d43ba1d20c067115c63db8f87982937f03a141ce1a2" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538455 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="caa19051a64d2139ffcaae693702a9add912d133e2c6ad1aa105da66652c3dbf" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538464 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="a397709beb06ddc2e3731ee791524a484e7b3ac4a47a134ded55ece54ff1c7b7" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538472 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="96a774aea1b785c85e9e4db9fdf687a4a017ec236987833423a664e390b9a635" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538119 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"328c19774f97092a3e3f498eec5b95de21679577a75dd683d879de5fa9cefbe6"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538528 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"06646bb3d2ff65eedaafa404d7d0aa7923592978978ae2271af7f090c46dc8fe"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538540 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"0321f9472e248b304e48ec950df8cd09d170ae9fdc38f48e874938cd6e895e85"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538551 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"46cbedf0abfd3f31c376d3f0fc737f9687de69d123ae655da80d749a3fc71176"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538561 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"13f1d2241b7ac6072e93443f2565800d1acf3174a4de209a3cdd382ce7b2c893"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538570 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"f6a023b7f054ee3eb28ad8d00763fbf33ab4b36409454d63e42ea87e4a6e617f"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538579 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"38075b27427df3883c9982c59fdeffdb66ca032b3966023e619377f2e77b8c9d"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538589 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"966a843ea31860db24f4b9279d146c5d253cc277653ba196f6ef2a6887360cb4"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538598 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"dcb1d92e14edd56619ba2577f2bd3ce1ada590038150f3acc568b1fa13bedae4"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538607 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"7f36f35de0e6d3a1b838b67e886b8a52e9398ce82c050e5b935cf0379ad62d86"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538617 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"6ab00368cc89fcf8161f0d43ba1d20c067115c63db8f87982937f03a141ce1a2"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538626 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"caa19051a64d2139ffcaae693702a9add912d133e2c6ad1aa105da66652c3dbf"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538635 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"a397709beb06ddc2e3731ee791524a484e7b3ac4a47a134ded55ece54ff1c7b7"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.538644 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"96a774aea1b785c85e9e4db9fdf687a4a017ec236987833423a664e390b9a635"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544511 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544536 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544544 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544553 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544561 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544570 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544577 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544585 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544593 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544602 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544610 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544618 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544625 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544632 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544668 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544688 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544699 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544709 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544719 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544730 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544740 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544750 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544759 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544768 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544776 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544784 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544793 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.544801 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.546768 4692 generic.go:334] "Generic (PLEG): container finished" podID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerID="d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.546814 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" event={"ID":"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a","Type":"ContainerDied","Data":"d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552862 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="140a3b0ef60391e40a7e1b3dbe1d9e95bbabab2391fdaa782882ae34c7dade4a" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552915 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="5daee84511b7c0cbba039fd3cbcf9d60634e7a175565d403aa95d2d6bc919b4c" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552925 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="b3f6440db1f20d77b29490fabaa341162a2757de5636fde37edaedeb487ba9e8" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552932 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="099b4f3feffdd401f91b122c7eebf222dc20cb7003286e6b3fdbb5b1dd988b2b" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552940 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="171d835038bf3f61e38646d3ac1a693d257467e87de91754cb63ad379961e59e" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552946 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="26c226cab26d728919ade2a94001899645db4128cc63122ef0b7e131dafe18e4" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552953 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="0574622c3ea59457671ae2d3e5912c7907129a0385b9617eb6c635283f6aedf8" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552961 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="4dcd159606ea3eb0a56b35c2bbc8955b8cf4ee1aeaddc4a57f92fb1b5c579662" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552969 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="dfc762ad83fc57e117251ed04fa58e82ee932457ae195cc07a45d8582c6f48b3" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552976 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="7d475f88fe15d82cbca79a9f0f44b9547c75bd6a08ec54a080ad68c5c9eb1667" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552984 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="704ea831492a29f5a461f465943223a5cf5a1275ba375cc94f8c0af1c8c9c7dc" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552990 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="ccb5034272c4443a5d30c3e77c9d8499de513d71548f5dfc4aa471c4cf783678" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.552997 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="d9fca122f4493bee53c9b8f344da247b633bac2de061a56c8724e98fc1516460" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553003 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="cbf6beff918423c3958f88e8749e875ff0a036cc342ad42299854971e4745a5a" exitCode=0 Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553221 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"140a3b0ef60391e40a7e1b3dbe1d9e95bbabab2391fdaa782882ae34c7dade4a"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553243 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"5daee84511b7c0cbba039fd3cbcf9d60634e7a175565d403aa95d2d6bc919b4c"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553254 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"b3f6440db1f20d77b29490fabaa341162a2757de5636fde37edaedeb487ba9e8"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553263 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"099b4f3feffdd401f91b122c7eebf222dc20cb7003286e6b3fdbb5b1dd988b2b"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553274 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"171d835038bf3f61e38646d3ac1a693d257467e87de91754cb63ad379961e59e"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553312 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"26c226cab26d728919ade2a94001899645db4128cc63122ef0b7e131dafe18e4"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553322 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"0574622c3ea59457671ae2d3e5912c7907129a0385b9617eb6c635283f6aedf8"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553332 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"4dcd159606ea3eb0a56b35c2bbc8955b8cf4ee1aeaddc4a57f92fb1b5c579662"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553343 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"dfc762ad83fc57e117251ed04fa58e82ee932457ae195cc07a45d8582c6f48b3"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553355 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"7d475f88fe15d82cbca79a9f0f44b9547c75bd6a08ec54a080ad68c5c9eb1667"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553369 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"704ea831492a29f5a461f465943223a5cf5a1275ba375cc94f8c0af1c8c9c7dc"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553381 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"ccb5034272c4443a5d30c3e77c9d8499de513d71548f5dfc4aa471c4cf783678"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553393 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"d9fca122f4493bee53c9b8f344da247b633bac2de061a56c8724e98fc1516460"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.553406 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"cbf6beff918423c3958f88e8749e875ff0a036cc342ad42299854971e4745a5a"} Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.950065 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:50 crc kubenswrapper[4692]: I0309 09:39:50.957193 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082189 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-log-httpd\") pod \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082254 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-scripts\") pod \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082329 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-ring-data-devices\") pod \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082577 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdqxz\" (UniqueName: \"kubernetes.io/projected/79ba58b3-a4a4-4375-b1e2-285f952e08a4-kube-api-access-zdqxz\") pod \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082615 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-dispersionconf\") pod \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082643 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gstn\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-kube-api-access-7gstn\") pod \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082666 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-swiftconf\") pod \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082702 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79ba58b3-a4a4-4375-b1e2-285f952e08a4-etc-swift\") pod \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\" (UID: \"79ba58b3-a4a4-4375-b1e2-285f952e08a4\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082744 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-config-data\") pod \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082789 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") pod \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.082829 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-run-httpd\") pod \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\" (UID: \"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a\") " Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.083420 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.083563 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "79ba58b3-a4a4-4375-b1e2-285f952e08a4" (UID: "79ba58b3-a4a4-4375-b1e2-285f952e08a4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.083977 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.084486 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79ba58b3-a4a4-4375-b1e2-285f952e08a4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "79ba58b3-a4a4-4375-b1e2-285f952e08a4" (UID: "79ba58b3-a4a4-4375-b1e2-285f952e08a4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.088987 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-kube-api-access-7gstn" (OuterVolumeSpecName: "kube-api-access-7gstn") pod "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a"). InnerVolumeSpecName "kube-api-access-7gstn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.089475 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79ba58b3-a4a4-4375-b1e2-285f952e08a4-kube-api-access-zdqxz" (OuterVolumeSpecName: "kube-api-access-zdqxz") pod "79ba58b3-a4a4-4375-b1e2-285f952e08a4" (UID: "79ba58b3-a4a4-4375-b1e2-285f952e08a4"). InnerVolumeSpecName "kube-api-access-zdqxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.089883 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.109865 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "79ba58b3-a4a4-4375-b1e2-285f952e08a4" (UID: "79ba58b3-a4a4-4375-b1e2-285f952e08a4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.113339 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "79ba58b3-a4a4-4375-b1e2-285f952e08a4" (UID: "79ba58b3-a4a4-4375-b1e2-285f952e08a4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.113590 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-scripts" (OuterVolumeSpecName: "scripts") pod "79ba58b3-a4a4-4375-b1e2-285f952e08a4" (UID: "79ba58b3-a4a4-4375-b1e2-285f952e08a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.125335 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-config-data" (OuterVolumeSpecName: "config-data") pod "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" (UID: "1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184560 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdqxz\" (UniqueName: \"kubernetes.io/projected/79ba58b3-a4a4-4375-b1e2-285f952e08a4-kube-api-access-zdqxz\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184795 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184809 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gstn\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-kube-api-access-7gstn\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184822 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79ba58b3-a4a4-4375-b1e2-285f952e08a4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184830 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79ba58b3-a4a4-4375-b1e2-285f952e08a4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184839 4692 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184847 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184856 4692 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184864 4692 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184873 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.184881 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79ba58b3-a4a4-4375-b1e2-285f952e08a4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.564956 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-n4sfg" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.564960 4692 scope.go:117] "RemoveContainer" containerID="b38c6942b9a7bf311d47525695855d615ccb8e90ebcb397ff550fd91174281bc" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.567998 4692 generic.go:334] "Generic (PLEG): container finished" podID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerID="099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1" exitCode=0 Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.568042 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" event={"ID":"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a","Type":"ContainerDied","Data":"099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1"} Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.568084 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" event={"ID":"1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a","Type":"ContainerDied","Data":"4efb13ce1cc00883326f4aad8e1e720264d37cbe0a5d6006f1a9cf67646b1166"} Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.568090 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-kv9nf" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.605832 4692 scope.go:117] "RemoveContainer" containerID="099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.618148 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kv9nf"] Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.625319 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-kv9nf"] Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.628537 4692 scope.go:117] "RemoveContainer" containerID="d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.645150 4692 scope.go:117] "RemoveContainer" containerID="099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1" Mar 09 09:39:51 crc kubenswrapper[4692]: E0309 09:39:51.645597 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1\": container with ID starting with 099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1 not found: ID does not exist" containerID="099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.645630 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1"} err="failed to get container status \"099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1\": rpc error: code = NotFound desc = could not find container \"099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1\": container with ID starting with 099b6d8abc9534da12713c2c2faf59a8489b265cd507df72228eadbd8d6432d1 not found: ID does not exist" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.645657 4692 scope.go:117] "RemoveContainer" containerID="d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4" Mar 09 09:39:51 crc kubenswrapper[4692]: E0309 09:39:51.645899 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4\": container with ID starting with d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4 not found: ID does not exist" containerID="d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4" Mar 09 09:39:51 crc kubenswrapper[4692]: I0309 09:39:51.645919 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4"} err="failed to get container status \"d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4\": rpc error: code = NotFound desc = could not find container \"d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4\": container with ID starting with d7ba9c070bc613fef0bc7790a3eff47e441808f8cce1aee7157e0294ad8a50d4 not found: ID does not exist" Mar 09 09:39:52 crc kubenswrapper[4692]: I0309 09:39:52.081020 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" path="/var/lib/kubelet/pods/1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a/volumes" Mar 09 09:39:52 crc kubenswrapper[4692]: I0309 09:39:52.081883 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79ba58b3-a4a4-4375-b1e2-285f952e08a4" path="/var/lib/kubelet/pods/79ba58b3-a4a4-4375-b1e2-285f952e08a4/volumes" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.137530 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550820-jnhqp"] Mar 09 09:40:00 crc kubenswrapper[4692]: E0309 09:40:00.142696 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79ba58b3-a4a4-4375-b1e2-285f952e08a4" containerName="swift-ring-rebalance" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.142719 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="79ba58b3-a4a4-4375-b1e2-285f952e08a4" containerName="swift-ring-rebalance" Mar 09 09:40:00 crc kubenswrapper[4692]: E0309 09:40:00.142733 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-server" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.142743 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-server" Mar 09 09:40:00 crc kubenswrapper[4692]: E0309 09:40:00.142769 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-httpd" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.142776 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-httpd" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.142950 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-server" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.142969 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e99c1fd-25a0-4f63-9eab-b9dddcbfc39a" containerName="proxy-httpd" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.142986 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="79ba58b3-a4a4-4375-b1e2-285f952e08a4" containerName="swift-ring-rebalance" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.143649 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550820-jnhqp" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.146536 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.146863 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.146868 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.149066 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550820-jnhqp"] Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.207688 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmfs5\" (UniqueName: \"kubernetes.io/projected/4193ac81-13f3-42db-afd3-bbc890b0000f-kube-api-access-qmfs5\") pod \"auto-csr-approver-29550820-jnhqp\" (UID: \"4193ac81-13f3-42db-afd3-bbc890b0000f\") " pod="openshift-infra/auto-csr-approver-29550820-jnhqp" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.309425 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmfs5\" (UniqueName: \"kubernetes.io/projected/4193ac81-13f3-42db-afd3-bbc890b0000f-kube-api-access-qmfs5\") pod \"auto-csr-approver-29550820-jnhqp\" (UID: \"4193ac81-13f3-42db-afd3-bbc890b0000f\") " pod="openshift-infra/auto-csr-approver-29550820-jnhqp" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.326084 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmfs5\" (UniqueName: \"kubernetes.io/projected/4193ac81-13f3-42db-afd3-bbc890b0000f-kube-api-access-qmfs5\") pod \"auto-csr-approver-29550820-jnhqp\" (UID: \"4193ac81-13f3-42db-afd3-bbc890b0000f\") " pod="openshift-infra/auto-csr-approver-29550820-jnhqp" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.479854 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550820-jnhqp" Mar 09 09:40:00 crc kubenswrapper[4692]: I0309 09:40:00.898123 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550820-jnhqp"] Mar 09 09:40:01 crc kubenswrapper[4692]: I0309 09:40:01.642121 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550820-jnhqp" event={"ID":"4193ac81-13f3-42db-afd3-bbc890b0000f","Type":"ContainerStarted","Data":"129e9d778805e2db2faeba1e35e4dfcdb8b9a023dcf5fbffb078ac2762f75298"} Mar 09 09:40:03 crc kubenswrapper[4692]: I0309 09:40:03.659425 4692 generic.go:334] "Generic (PLEG): container finished" podID="4193ac81-13f3-42db-afd3-bbc890b0000f" containerID="5ee28c43242d8b55f00a22bcffc9a222f8c0cec8d60d2e6f93410be897772cfb" exitCode=0 Mar 09 09:40:03 crc kubenswrapper[4692]: I0309 09:40:03.659538 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550820-jnhqp" event={"ID":"4193ac81-13f3-42db-afd3-bbc890b0000f","Type":"ContainerDied","Data":"5ee28c43242d8b55f00a22bcffc9a222f8c0cec8d60d2e6f93410be897772cfb"} Mar 09 09:40:04 crc kubenswrapper[4692]: I0309 09:40:04.967676 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550820-jnhqp" Mar 09 09:40:05 crc kubenswrapper[4692]: I0309 09:40:05.073784 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmfs5\" (UniqueName: \"kubernetes.io/projected/4193ac81-13f3-42db-afd3-bbc890b0000f-kube-api-access-qmfs5\") pod \"4193ac81-13f3-42db-afd3-bbc890b0000f\" (UID: \"4193ac81-13f3-42db-afd3-bbc890b0000f\") " Mar 09 09:40:05 crc kubenswrapper[4692]: I0309 09:40:05.079266 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4193ac81-13f3-42db-afd3-bbc890b0000f-kube-api-access-qmfs5" (OuterVolumeSpecName: "kube-api-access-qmfs5") pod "4193ac81-13f3-42db-afd3-bbc890b0000f" (UID: "4193ac81-13f3-42db-afd3-bbc890b0000f"). InnerVolumeSpecName "kube-api-access-qmfs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:05 crc kubenswrapper[4692]: I0309 09:40:05.175775 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmfs5\" (UniqueName: \"kubernetes.io/projected/4193ac81-13f3-42db-afd3-bbc890b0000f-kube-api-access-qmfs5\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:05 crc kubenswrapper[4692]: I0309 09:40:05.676711 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550820-jnhqp" event={"ID":"4193ac81-13f3-42db-afd3-bbc890b0000f","Type":"ContainerDied","Data":"129e9d778805e2db2faeba1e35e4dfcdb8b9a023dcf5fbffb078ac2762f75298"} Mar 09 09:40:05 crc kubenswrapper[4692]: I0309 09:40:05.677106 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="129e9d778805e2db2faeba1e35e4dfcdb8b9a023dcf5fbffb078ac2762f75298" Mar 09 09:40:05 crc kubenswrapper[4692]: I0309 09:40:05.676756 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550820-jnhqp" Mar 09 09:40:06 crc kubenswrapper[4692]: I0309 09:40:06.045335 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550814-wj7qv"] Mar 09 09:40:06 crc kubenswrapper[4692]: I0309 09:40:06.051177 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550814-wj7qv"] Mar 09 09:40:06 crc kubenswrapper[4692]: I0309 09:40:06.079915 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9c73711-7e88-4f9a-a2fc-3ace9e4144ec" path="/var/lib/kubelet/pods/f9c73711-7e88-4f9a-a2fc-3ace9e4144ec/volumes" Mar 09 09:40:19 crc kubenswrapper[4692]: I0309 09:40:19.788562 4692 generic.go:334] "Generic (PLEG): container finished" podID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerID="280979651d624f8b196a49bb676333d78dd6e3ba38a2c3044d7159baa49a045b" exitCode=137 Mar 09 09:40:19 crc kubenswrapper[4692]: I0309 09:40:19.788635 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"280979651d624f8b196a49bb676333d78dd6e3ba38a2c3044d7159baa49a045b"} Mar 09 09:40:19 crc kubenswrapper[4692]: E0309 09:40:19.885356 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda75a44c6_af49_4a59_a898_14b90d1eaba9.slice/crio-conmon-75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda75a44c6_af49_4a59_a898_14b90d1eaba9.slice/crio-75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d.scope\": RecentStats: unable to find data in memory cache]" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.104145 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.200123 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-etc-swift\") pod \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.200427 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrfvn\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-kube-api-access-zrfvn\") pod \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.200458 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.200498 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-lock\") pod \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.200545 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-cache\") pod \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\" (UID: \"6904b3a2-2562-46b8-8c7c-634d9c7f917e\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.201307 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-cache" (OuterVolumeSpecName: "cache") pod "6904b3a2-2562-46b8-8c7c-634d9c7f917e" (UID: "6904b3a2-2562-46b8-8c7c-634d9c7f917e"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.201942 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-lock" (OuterVolumeSpecName: "lock") pod "6904b3a2-2562-46b8-8c7c-634d9c7f917e" (UID: "6904b3a2-2562-46b8-8c7c-634d9c7f917e"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.202188 4692 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-lock\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.202211 4692 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6904b3a2-2562-46b8-8c7c-634d9c7f917e-cache\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.206634 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "swift") pod "6904b3a2-2562-46b8-8c7c-634d9c7f917e" (UID: "6904b3a2-2562-46b8-8c7c-634d9c7f917e"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.206900 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6904b3a2-2562-46b8-8c7c-634d9c7f917e" (UID: "6904b3a2-2562-46b8-8c7c-634d9c7f917e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.208275 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-kube-api-access-zrfvn" (OuterVolumeSpecName: "kube-api-access-zrfvn") pod "6904b3a2-2562-46b8-8c7c-634d9c7f917e" (UID: "6904b3a2-2562-46b8-8c7c-634d9c7f917e"). InnerVolumeSpecName "kube-api-access-zrfvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.208806 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.214635 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.302949 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"243cb7f7-785c-4681-b344-ea4073562c84\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303081 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") pod \"243cb7f7-785c-4681-b344-ea4073562c84\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303109 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z925c\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-kube-api-access-z925c\") pod \"a75a44c6-af49-4a59-a898-14b90d1eaba9\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303610 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-etc-swift\") pod \"a75a44c6-af49-4a59-a898-14b90d1eaba9\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303649 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-cache\") pod \"243cb7f7-785c-4681-b344-ea4073562c84\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303701 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-lock\") pod \"243cb7f7-785c-4681-b344-ea4073562c84\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303731 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8hp6\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-kube-api-access-t8hp6\") pod \"243cb7f7-785c-4681-b344-ea4073562c84\" (UID: \"243cb7f7-785c-4681-b344-ea4073562c84\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303757 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"a75a44c6-af49-4a59-a898-14b90d1eaba9\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303799 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-lock\") pod \"a75a44c6-af49-4a59-a898-14b90d1eaba9\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.303824 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-cache\") pod \"a75a44c6-af49-4a59-a898-14b90d1eaba9\" (UID: \"a75a44c6-af49-4a59-a898-14b90d1eaba9\") " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.304238 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-lock" (OuterVolumeSpecName: "lock") pod "243cb7f7-785c-4681-b344-ea4073562c84" (UID: "243cb7f7-785c-4681-b344-ea4073562c84"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.304367 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-cache" (OuterVolumeSpecName: "cache") pod "243cb7f7-785c-4681-b344-ea4073562c84" (UID: "243cb7f7-785c-4681-b344-ea4073562c84"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.304565 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-lock" (OuterVolumeSpecName: "lock") pod "a75a44c6-af49-4a59-a898-14b90d1eaba9" (UID: "a75a44c6-af49-4a59-a898-14b90d1eaba9"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.304626 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.304721 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-cache" (OuterVolumeSpecName: "cache") pod "a75a44c6-af49-4a59-a898-14b90d1eaba9" (UID: "a75a44c6-af49-4a59-a898-14b90d1eaba9"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.304859 4692 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-cache\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.304939 4692 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/243cb7f7-785c-4681-b344-ea4073562c84-lock\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.305176 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrfvn\" (UniqueName: \"kubernetes.io/projected/6904b3a2-2562-46b8-8c7c-634d9c7f917e-kube-api-access-zrfvn\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.305295 4692 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.305663 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "swift") pod "243cb7f7-785c-4681-b344-ea4073562c84" (UID: "243cb7f7-785c-4681-b344-ea4073562c84"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.306442 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "243cb7f7-785c-4681-b344-ea4073562c84" (UID: "243cb7f7-785c-4681-b344-ea4073562c84"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.306868 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-kube-api-access-t8hp6" (OuterVolumeSpecName: "kube-api-access-t8hp6") pod "243cb7f7-785c-4681-b344-ea4073562c84" (UID: "243cb7f7-785c-4681-b344-ea4073562c84"). InnerVolumeSpecName "kube-api-access-t8hp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.306875 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-kube-api-access-z925c" (OuterVolumeSpecName: "kube-api-access-z925c") pod "a75a44c6-af49-4a59-a898-14b90d1eaba9" (UID: "a75a44c6-af49-4a59-a898-14b90d1eaba9"). InnerVolumeSpecName "kube-api-access-z925c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.307301 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a75a44c6-af49-4a59-a898-14b90d1eaba9" (UID: "a75a44c6-af49-4a59-a898-14b90d1eaba9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.308189 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "swift") pod "a75a44c6-af49-4a59-a898-14b90d1eaba9" (UID: "a75a44c6-af49-4a59-a898-14b90d1eaba9"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.318636 4692 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.406912 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.406975 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z925c\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-kube-api-access-z925c\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.406986 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a75a44c6-af49-4a59-a898-14b90d1eaba9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.406995 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8hp6\" (UniqueName: \"kubernetes.io/projected/243cb7f7-785c-4681-b344-ea4073562c84-kube-api-access-t8hp6\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.407005 4692 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.407059 4692 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.407070 4692 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-lock\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.407080 4692 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a75a44c6-af49-4a59-a898-14b90d1eaba9-cache\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.407095 4692 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.418894 4692 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.419616 4692 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.508948 4692 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.508986 4692 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.805560 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"6904b3a2-2562-46b8-8c7c-634d9c7f917e","Type":"ContainerDied","Data":"81269b346e9df9f2f6656d2e4deab5af9b59db56c995ea774262cb484f0bb208"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.805636 4692 scope.go:117] "RemoveContainer" containerID="280979651d624f8b196a49bb676333d78dd6e3ba38a2c3044d7159baa49a045b" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.805926 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.816984 4692 generic.go:334] "Generic (PLEG): container finished" podID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerID="75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d" exitCode=137 Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.817069 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.817108 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"a75a44c6-af49-4a59-a898-14b90d1eaba9","Type":"ContainerDied","Data":"c8ae29f7cb9bb78b1d1be941154615915a0a3b76626caa822de4d652bb6226a0"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.817266 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835014 4692 generic.go:334] "Generic (PLEG): container finished" podID="243cb7f7-785c-4681-b344-ea4073562c84" containerID="86718cd93961aca462a530e4a4791cb7e77a447add0e2a72803c9f229c6ede23" exitCode=137 Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835097 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"86718cd93961aca462a530e4a4791cb7e77a447add0e2a72803c9f229c6ede23"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835186 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26c226cab26d728919ade2a94001899645db4128cc63122ef0b7e131dafe18e4"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835201 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0574622c3ea59457671ae2d3e5912c7907129a0385b9617eb6c635283f6aedf8"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835214 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4dcd159606ea3eb0a56b35c2bbc8955b8cf4ee1aeaddc4a57f92fb1b5c579662"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835220 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dfc762ad83fc57e117251ed04fa58e82ee932457ae195cc07a45d8582c6f48b3"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835226 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d475f88fe15d82cbca79a9f0f44b9547c75bd6a08ec54a080ad68c5c9eb1667"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835192 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835231 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"704ea831492a29f5a461f465943223a5cf5a1275ba375cc94f8c0af1c8c9c7dc"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835367 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccb5034272c4443a5d30c3e77c9d8499de513d71548f5dfc4aa471c4cf783678"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835392 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9fca122f4493bee53c9b8f344da247b633bac2de061a56c8724e98fc1516460"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835400 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cbf6beff918423c3958f88e8749e875ff0a036cc342ad42299854971e4745a5a"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835422 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"243cb7f7-785c-4681-b344-ea4073562c84","Type":"ContainerDied","Data":"9ff78cf169b13930889ded5fc90d081b4e3777328901c64edbbe2681eedc2153"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835447 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86718cd93961aca462a530e4a4791cb7e77a447add0e2a72803c9f229c6ede23"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835455 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"140a3b0ef60391e40a7e1b3dbe1d9e95bbabab2391fdaa782882ae34c7dade4a"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835462 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5daee84511b7c0cbba039fd3cbcf9d60634e7a175565d403aa95d2d6bc919b4c"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835469 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3f6440db1f20d77b29490fabaa341162a2757de5636fde37edaedeb487ba9e8"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835476 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"099b4f3feffdd401f91b122c7eebf222dc20cb7003286e6b3fdbb5b1dd988b2b"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835482 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"171d835038bf3f61e38646d3ac1a693d257467e87de91754cb63ad379961e59e"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835489 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"26c226cab26d728919ade2a94001899645db4128cc63122ef0b7e131dafe18e4"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835496 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0574622c3ea59457671ae2d3e5912c7907129a0385b9617eb6c635283f6aedf8"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835504 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4dcd159606ea3eb0a56b35c2bbc8955b8cf4ee1aeaddc4a57f92fb1b5c579662"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835510 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dfc762ad83fc57e117251ed04fa58e82ee932457ae195cc07a45d8582c6f48b3"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835516 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d475f88fe15d82cbca79a9f0f44b9547c75bd6a08ec54a080ad68c5c9eb1667"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835523 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"704ea831492a29f5a461f465943223a5cf5a1275ba375cc94f8c0af1c8c9c7dc"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835532 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ccb5034272c4443a5d30c3e77c9d8499de513d71548f5dfc4aa471c4cf783678"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835538 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d9fca122f4493bee53c9b8f344da247b633bac2de061a56c8724e98fc1516460"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.835548 4692 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cbf6beff918423c3958f88e8749e875ff0a036cc342ad42299854971e4745a5a"} Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.869964 4692 scope.go:117] "RemoveContainer" containerID="328c19774f97092a3e3f498eec5b95de21679577a75dd683d879de5fa9cefbe6" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.879222 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.889683 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.893745 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.906989 4692 scope.go:117] "RemoveContainer" containerID="06646bb3d2ff65eedaafa404d7d0aa7923592978978ae2271af7f090c46dc8fe" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.907040 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.910547 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.917244 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.925712 4692 scope.go:117] "RemoveContainer" containerID="0321f9472e248b304e48ec950df8cd09d170ae9fdc38f48e874938cd6e895e85" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.943080 4692 scope.go:117] "RemoveContainer" containerID="46cbedf0abfd3f31c376d3f0fc737f9687de69d123ae655da80d749a3fc71176" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.962003 4692 scope.go:117] "RemoveContainer" containerID="13f1d2241b7ac6072e93443f2565800d1acf3174a4de209a3cdd382ce7b2c893" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.979392 4692 scope.go:117] "RemoveContainer" containerID="f6a023b7f054ee3eb28ad8d00763fbf33ab4b36409454d63e42ea87e4a6e617f" Mar 09 09:40:20 crc kubenswrapper[4692]: I0309 09:40:20.995145 4692 scope.go:117] "RemoveContainer" containerID="38075b27427df3883c9982c59fdeffdb66ca032b3966023e619377f2e77b8c9d" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.011552 4692 scope.go:117] "RemoveContainer" containerID="966a843ea31860db24f4b9279d146c5d253cc277653ba196f6ef2a6887360cb4" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.026601 4692 scope.go:117] "RemoveContainer" containerID="dcb1d92e14edd56619ba2577f2bd3ce1ada590038150f3acc568b1fa13bedae4" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.042313 4692 scope.go:117] "RemoveContainer" containerID="7f36f35de0e6d3a1b838b67e886b8a52e9398ce82c050e5b935cf0379ad62d86" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.059179 4692 scope.go:117] "RemoveContainer" containerID="6ab00368cc89fcf8161f0d43ba1d20c067115c63db8f87982937f03a141ce1a2" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.078781 4692 scope.go:117] "RemoveContainer" containerID="caa19051a64d2139ffcaae693702a9add912d133e2c6ad1aa105da66652c3dbf" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.101663 4692 scope.go:117] "RemoveContainer" containerID="a397709beb06ddc2e3731ee791524a484e7b3ac4a47a134ded55ece54ff1c7b7" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.119134 4692 scope.go:117] "RemoveContainer" containerID="96a774aea1b785c85e9e4db9fdf687a4a017ec236987833423a664e390b9a635" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.135219 4692 scope.go:117] "RemoveContainer" containerID="75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.149499 4692 scope.go:117] "RemoveContainer" containerID="2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.166067 4692 scope.go:117] "RemoveContainer" containerID="ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.180982 4692 scope.go:117] "RemoveContainer" containerID="dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.195384 4692 scope.go:117] "RemoveContainer" containerID="5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.213008 4692 scope.go:117] "RemoveContainer" containerID="c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.229607 4692 scope.go:117] "RemoveContainer" containerID="1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.244478 4692 scope.go:117] "RemoveContainer" containerID="a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.260605 4692 scope.go:117] "RemoveContainer" containerID="24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.276678 4692 scope.go:117] "RemoveContainer" containerID="68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.301408 4692 scope.go:117] "RemoveContainer" containerID="53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.323366 4692 scope.go:117] "RemoveContainer" containerID="ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.339222 4692 scope.go:117] "RemoveContainer" containerID="8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.355944 4692 scope.go:117] "RemoveContainer" containerID="462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.374334 4692 scope.go:117] "RemoveContainer" containerID="8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.391615 4692 scope.go:117] "RemoveContainer" containerID="75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.392132 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d\": container with ID starting with 75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d not found: ID does not exist" containerID="75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.392200 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d"} err="failed to get container status \"75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d\": rpc error: code = NotFound desc = could not find container \"75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d\": container with ID starting with 75f0fc873ec45e0837aa4315dd6e32463299257b07f790dcdbd2a6829fd4b46d not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.392222 4692 scope.go:117] "RemoveContainer" containerID="2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.392561 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba\": container with ID starting with 2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba not found: ID does not exist" containerID="2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.392609 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba"} err="failed to get container status \"2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba\": rpc error: code = NotFound desc = could not find container \"2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba\": container with ID starting with 2282e079cfd2420f3357e624fdfa98bb0a78f1e2d6698bf175f195e416532bba not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.392644 4692 scope.go:117] "RemoveContainer" containerID="ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.392932 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2\": container with ID starting with ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2 not found: ID does not exist" containerID="ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.392959 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2"} err="failed to get container status \"ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2\": rpc error: code = NotFound desc = could not find container \"ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2\": container with ID starting with ba1ad2f83cb7011ec5340cb6bf173606db2727d6ecb2621acfafa6ef285de0f2 not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.392975 4692 scope.go:117] "RemoveContainer" containerID="dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.393215 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc\": container with ID starting with dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc not found: ID does not exist" containerID="dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.393281 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc"} err="failed to get container status \"dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc\": rpc error: code = NotFound desc = could not find container \"dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc\": container with ID starting with dfbd152788e112184312e192f6389afaf0572c5deeb66c92ee1059555d6fcedc not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.393297 4692 scope.go:117] "RemoveContainer" containerID="5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.394651 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016\": container with ID starting with 5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016 not found: ID does not exist" containerID="5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.394681 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016"} err="failed to get container status \"5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016\": rpc error: code = NotFound desc = could not find container \"5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016\": container with ID starting with 5e3e3bcf4f5509fa783dc0d7583bb93539a1454c1864f457a52fe5cedce3e016 not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.394703 4692 scope.go:117] "RemoveContainer" containerID="c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.394965 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c\": container with ID starting with c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c not found: ID does not exist" containerID="c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.394987 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c"} err="failed to get container status \"c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c\": rpc error: code = NotFound desc = could not find container \"c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c\": container with ID starting with c6db20028d38dfaefa316e254d6bcbb903c97a72b288e69be2131ae4f14bdb0c not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.395005 4692 scope.go:117] "RemoveContainer" containerID="1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.395284 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c\": container with ID starting with 1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c not found: ID does not exist" containerID="1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.395314 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c"} err="failed to get container status \"1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c\": rpc error: code = NotFound desc = could not find container \"1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c\": container with ID starting with 1eae5cb4901ffd367039ebf9afc554af01d61e961da093c1985a2664afadae6c not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.395331 4692 scope.go:117] "RemoveContainer" containerID="a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.395733 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d\": container with ID starting with a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d not found: ID does not exist" containerID="a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.395774 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d"} err="failed to get container status \"a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d\": rpc error: code = NotFound desc = could not find container \"a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d\": container with ID starting with a18878179247976a7a9fc4089e2bf6e80c0acba7b2bdf5236326501398c0b88d not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.395812 4692 scope.go:117] "RemoveContainer" containerID="24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.396096 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8\": container with ID starting with 24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8 not found: ID does not exist" containerID="24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.396119 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8"} err="failed to get container status \"24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8\": rpc error: code = NotFound desc = could not find container \"24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8\": container with ID starting with 24c7fe3a7c9f788cb340fdab399cab6fa01c95c49a309e63e04399793431bba8 not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.396138 4692 scope.go:117] "RemoveContainer" containerID="68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.396592 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d\": container with ID starting with 68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d not found: ID does not exist" containerID="68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.396644 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d"} err="failed to get container status \"68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d\": rpc error: code = NotFound desc = could not find container \"68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d\": container with ID starting with 68be0f91de4603da2b3c6cbd326e015298b5330bcb976a71198fc8b27a7fd39d not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.396688 4692 scope.go:117] "RemoveContainer" containerID="53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.397035 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11\": container with ID starting with 53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11 not found: ID does not exist" containerID="53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.397060 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11"} err="failed to get container status \"53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11\": rpc error: code = NotFound desc = could not find container \"53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11\": container with ID starting with 53207b5ebaa650ef2c043e6834688cc8496d3d55c8ebf59a2d56047991c9dd11 not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.397074 4692 scope.go:117] "RemoveContainer" containerID="ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.397514 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e\": container with ID starting with ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e not found: ID does not exist" containerID="ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.397537 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e"} err="failed to get container status \"ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e\": rpc error: code = NotFound desc = could not find container \"ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e\": container with ID starting with ad72852db832166230ed656b81cf7f8351b81ecbfb18a8eb562829530a4a236e not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.397559 4692 scope.go:117] "RemoveContainer" containerID="8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.397917 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747\": container with ID starting with 8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747 not found: ID does not exist" containerID="8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.397936 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747"} err="failed to get container status \"8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747\": rpc error: code = NotFound desc = could not find container \"8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747\": container with ID starting with 8e04ce970c71eb930bbe87975e9e8151aec4b5064971bc6d3436c55a00cb2747 not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.397948 4692 scope.go:117] "RemoveContainer" containerID="462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.398220 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4\": container with ID starting with 462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4 not found: ID does not exist" containerID="462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.398261 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4"} err="failed to get container status \"462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4\": rpc error: code = NotFound desc = could not find container \"462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4\": container with ID starting with 462534d076fb803c2b7b1444b26f8a42177541e94d6508ce5fdec238013a5cc4 not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.398310 4692 scope.go:117] "RemoveContainer" containerID="8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796" Mar 09 09:40:21 crc kubenswrapper[4692]: E0309 09:40:21.398668 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796\": container with ID starting with 8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796 not found: ID does not exist" containerID="8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.398708 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796"} err="failed to get container status \"8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796\": rpc error: code = NotFound desc = could not find container \"8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796\": container with ID starting with 8b2f305d231be1337f4ac54614f675e4db1dbf5cc5cc001ccbde59a4531a7796 not found: ID does not exist" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.398726 4692 scope.go:117] "RemoveContainer" containerID="86718cd93961aca462a530e4a4791cb7e77a447add0e2a72803c9f229c6ede23" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.413127 4692 scope.go:117] "RemoveContainer" containerID="140a3b0ef60391e40a7e1b3dbe1d9e95bbabab2391fdaa782882ae34c7dade4a" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.431683 4692 scope.go:117] "RemoveContainer" containerID="5daee84511b7c0cbba039fd3cbcf9d60634e7a175565d403aa95d2d6bc919b4c" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.445521 4692 scope.go:117] "RemoveContainer" containerID="b3f6440db1f20d77b29490fabaa341162a2757de5636fde37edaedeb487ba9e8" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.459899 4692 scope.go:117] "RemoveContainer" containerID="099b4f3feffdd401f91b122c7eebf222dc20cb7003286e6b3fdbb5b1dd988b2b" Mar 09 09:40:21 crc kubenswrapper[4692]: I0309 09:40:21.473125 4692 scope.go:117] "RemoveContainer" containerID="171d835038bf3f61e38646d3ac1a693d257467e87de91754cb63ad379961e59e" Mar 09 09:40:22 crc kubenswrapper[4692]: I0309 09:40:22.081210 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="243cb7f7-785c-4681-b344-ea4073562c84" path="/var/lib/kubelet/pods/243cb7f7-785c-4681-b344-ea4073562c84/volumes" Mar 09 09:40:22 crc kubenswrapper[4692]: I0309 09:40:22.083144 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" path="/var/lib/kubelet/pods/6904b3a2-2562-46b8-8c7c-634d9c7f917e/volumes" Mar 09 09:40:22 crc kubenswrapper[4692]: I0309 09:40:22.085397 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" path="/var/lib/kubelet/pods/a75a44c6-af49-4a59-a898-14b90d1eaba9/volumes" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.681892 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682642 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682655 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682666 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682671 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682680 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682686 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682693 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682699 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682712 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682720 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682728 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682735 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682746 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682753 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682766 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682774 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682785 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682791 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682819 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682827 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682837 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682842 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682851 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682857 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682865 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682871 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682881 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682886 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682895 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682902 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682912 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682917 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682927 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682934 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682940 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682945 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682954 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682959 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682966 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682971 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682980 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.682985 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.682996 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683001 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683010 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683016 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683026 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683031 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683041 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683046 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683056 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683062 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683072 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683077 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683087 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683093 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683101 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683106 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683115 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683120 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683130 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683136 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683144 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683150 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683176 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683184 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683196 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683203 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683209 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683215 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683224 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683229 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683239 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683245 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683254 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683260 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683266 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683271 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683279 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683284 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683291 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683297 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683304 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683310 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683317 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683322 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683331 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683337 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683347 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683352 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.683359 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4193ac81-13f3-42db-afd3-bbc890b0000f" containerName="oc" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683366 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4193ac81-13f3-42db-afd3-bbc890b0000f" containerName="oc" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683477 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683487 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683494 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683501 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683510 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683517 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683526 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683534 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683539 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683545 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683554 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683563 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683570 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683578 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-expirer" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683587 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683593 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683603 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683611 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683620 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683629 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683637 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4193ac81-13f3-42db-afd3-bbc890b0000f" containerName="oc" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683645 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683652 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683662 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683668 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683675 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683681 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683690 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="rsync" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683696 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683705 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683713 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683721 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683729 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683737 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683742 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="object-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683749 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="container-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683756 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683764 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="object-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683771 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="account-replicator" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683779 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683785 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683791 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="account-auditor" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683799 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-server" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683807 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="243cb7f7-785c-4681-b344-ea4073562c84" containerName="swift-recon-cron" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683814 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6904b3a2-2562-46b8-8c7c-634d9c7f917e" containerName="account-reaper" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.683821 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75a44c6-af49-4a59-a898-14b90d1eaba9" containerName="container-updater" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.688145 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.692328 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-files" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.692530 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-conf" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.692670 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-storage-config-data" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.694642 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-swift-dockercfg-x8krb" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.704419 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.760018 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lzbn\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-kube-api-access-4lzbn\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.760072 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-lock\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.760095 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.760239 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.760431 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-cache\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.861205 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-cache\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.861256 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lzbn\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-kube-api-access-4lzbn\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.861279 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-lock\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.861304 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.861371 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.861509 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.861522 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:40:23 crc kubenswrapper[4692]: E0309 09:40:23.861565 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift podName:c01b6df5-bc9c-4abd-9ce5-e928b06580c7 nodeName:}" failed. No retries permitted until 2026-03-09 09:40:24.361549531 +0000 UTC m=+1225.186285112 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift") pod "swift-storage-0" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7") : configmap "swift-ring-files" not found Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.862201 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") device mount path \"/mnt/openstack/pv03\"" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.862270 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-lock\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.862535 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-cache\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.880517 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lzbn\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-kube-api-access-4lzbn\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:23 crc kubenswrapper[4692]: I0309 09:40:23.881560 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.081068 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-knbvh"] Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.081795 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.083287 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-proxy-config-data" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.083411 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.084389 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.091716 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-knbvh"] Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.165734 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70cc1105-fd62-42e9-b6cb-612adea92724-etc-swift\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.166077 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-dispersionconf\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.166207 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-swiftconf\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.166344 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp87l\" (UniqueName: \"kubernetes.io/projected/70cc1105-fd62-42e9-b6cb-612adea92724-kube-api-access-rp87l\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.166761 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-ring-data-devices\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.166876 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-scripts\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.268132 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-dispersionconf\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.268256 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-swiftconf\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.268292 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp87l\" (UniqueName: \"kubernetes.io/projected/70cc1105-fd62-42e9-b6cb-612adea92724-kube-api-access-rp87l\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.268944 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-ring-data-devices\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.268985 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-scripts\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.269048 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70cc1105-fd62-42e9-b6cb-612adea92724-etc-swift\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.269400 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70cc1105-fd62-42e9-b6cb-612adea92724-etc-swift\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.269570 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-ring-data-devices\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.269896 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-scripts\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.272871 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-dispersionconf\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.282127 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-swiftconf\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.285408 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp87l\" (UniqueName: \"kubernetes.io/projected/70cc1105-fd62-42e9-b6cb-612adea92724-kube-api-access-rp87l\") pod \"swift-ring-rebalance-knbvh\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.370647 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:24 crc kubenswrapper[4692]: E0309 09:40:24.370813 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:40:24 crc kubenswrapper[4692]: E0309 09:40:24.370830 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:40:24 crc kubenswrapper[4692]: E0309 09:40:24.370888 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift podName:c01b6df5-bc9c-4abd-9ce5-e928b06580c7 nodeName:}" failed. No retries permitted until 2026-03-09 09:40:25.37087208 +0000 UTC m=+1226.195607661 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift") pod "swift-storage-0" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7") : configmap "swift-ring-files" not found Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.403696 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.855452 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-knbvh"] Mar 09 09:40:24 crc kubenswrapper[4692]: I0309 09:40:24.883879 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" event={"ID":"70cc1105-fd62-42e9-b6cb-612adea92724","Type":"ContainerStarted","Data":"6c58c661af0e375467a4809e919e99a9232be284879feafb6a0c88946e5c128a"} Mar 09 09:40:25 crc kubenswrapper[4692]: I0309 09:40:25.386100 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:25 crc kubenswrapper[4692]: E0309 09:40:25.386281 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:40:25 crc kubenswrapper[4692]: E0309 09:40:25.386521 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:40:25 crc kubenswrapper[4692]: E0309 09:40:25.386576 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift podName:c01b6df5-bc9c-4abd-9ce5-e928b06580c7 nodeName:}" failed. No retries permitted until 2026-03-09 09:40:27.38655979 +0000 UTC m=+1228.211295371 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift") pod "swift-storage-0" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7") : configmap "swift-ring-files" not found Mar 09 09:40:25 crc kubenswrapper[4692]: I0309 09:40:25.891645 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" event={"ID":"70cc1105-fd62-42e9-b6cb-612adea92724","Type":"ContainerStarted","Data":"c08c64838317ffe2b9a3715a7ef725eeeed77e577e7eae85c70ab5e550be268b"} Mar 09 09:40:25 crc kubenswrapper[4692]: I0309 09:40:25.909376 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" podStartSLOduration=1.909354703 podStartE2EDuration="1.909354703s" podCreationTimestamp="2026-03-09 09:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:40:25.904608182 +0000 UTC m=+1226.729343773" watchObservedRunningTime="2026-03-09 09:40:25.909354703 +0000 UTC m=+1226.734090284" Mar 09 09:40:27 crc kubenswrapper[4692]: I0309 09:40:27.419235 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:27 crc kubenswrapper[4692]: E0309 09:40:27.419739 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:40:27 crc kubenswrapper[4692]: E0309 09:40:27.419756 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:40:27 crc kubenswrapper[4692]: E0309 09:40:27.419804 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift podName:c01b6df5-bc9c-4abd-9ce5-e928b06580c7 nodeName:}" failed. No retries permitted until 2026-03-09 09:40:31.419787821 +0000 UTC m=+1232.244523402 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift") pod "swift-storage-0" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7") : configmap "swift-ring-files" not found Mar 09 09:40:31 crc kubenswrapper[4692]: I0309 09:40:31.488987 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:31 crc kubenswrapper[4692]: E0309 09:40:31.489253 4692 projected.go:288] Couldn't get configMap swift-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Mar 09 09:40:31 crc kubenswrapper[4692]: E0309 09:40:31.490961 4692 projected.go:194] Error preparing data for projected volume etc-swift for pod swift-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Mar 09 09:40:31 crc kubenswrapper[4692]: E0309 09:40:31.491052 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift podName:c01b6df5-bc9c-4abd-9ce5-e928b06580c7 nodeName:}" failed. No retries permitted until 2026-03-09 09:40:39.491028105 +0000 UTC m=+1240.315763686 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift") pod "swift-storage-0" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7") : configmap "swift-ring-files" not found Mar 09 09:40:31 crc kubenswrapper[4692]: I0309 09:40:31.940506 4692 generic.go:334] "Generic (PLEG): container finished" podID="70cc1105-fd62-42e9-b6cb-612adea92724" containerID="c08c64838317ffe2b9a3715a7ef725eeeed77e577e7eae85c70ab5e550be268b" exitCode=0 Mar 09 09:40:31 crc kubenswrapper[4692]: I0309 09:40:31.940550 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" event={"ID":"70cc1105-fd62-42e9-b6cb-612adea92724","Type":"ContainerDied","Data":"c08c64838317ffe2b9a3715a7ef725eeeed77e577e7eae85c70ab5e550be268b"} Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.240616 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.315490 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp87l\" (UniqueName: \"kubernetes.io/projected/70cc1105-fd62-42e9-b6cb-612adea92724-kube-api-access-rp87l\") pod \"70cc1105-fd62-42e9-b6cb-612adea92724\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.315574 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-scripts\") pod \"70cc1105-fd62-42e9-b6cb-612adea92724\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.315607 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-dispersionconf\") pod \"70cc1105-fd62-42e9-b6cb-612adea92724\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.315701 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-ring-data-devices\") pod \"70cc1105-fd62-42e9-b6cb-612adea92724\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.315735 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-swiftconf\") pod \"70cc1105-fd62-42e9-b6cb-612adea92724\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.315781 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70cc1105-fd62-42e9-b6cb-612adea92724-etc-swift\") pod \"70cc1105-fd62-42e9-b6cb-612adea92724\" (UID: \"70cc1105-fd62-42e9-b6cb-612adea92724\") " Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.317482 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70cc1105-fd62-42e9-b6cb-612adea92724-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "70cc1105-fd62-42e9-b6cb-612adea92724" (UID: "70cc1105-fd62-42e9-b6cb-612adea92724"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.317562 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "70cc1105-fd62-42e9-b6cb-612adea92724" (UID: "70cc1105-fd62-42e9-b6cb-612adea92724"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.322271 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70cc1105-fd62-42e9-b6cb-612adea92724-kube-api-access-rp87l" (OuterVolumeSpecName: "kube-api-access-rp87l") pod "70cc1105-fd62-42e9-b6cb-612adea92724" (UID: "70cc1105-fd62-42e9-b6cb-612adea92724"). InnerVolumeSpecName "kube-api-access-rp87l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.339016 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "70cc1105-fd62-42e9-b6cb-612adea92724" (UID: "70cc1105-fd62-42e9-b6cb-612adea92724"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.339471 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "70cc1105-fd62-42e9-b6cb-612adea92724" (UID: "70cc1105-fd62-42e9-b6cb-612adea92724"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.343504 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-scripts" (OuterVolumeSpecName: "scripts") pod "70cc1105-fd62-42e9-b6cb-612adea92724" (UID: "70cc1105-fd62-42e9-b6cb-612adea92724"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.418773 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.418818 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.418832 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/70cc1105-fd62-42e9-b6cb-612adea92724-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.418845 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/70cc1105-fd62-42e9-b6cb-612adea92724-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.418862 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp87l\" (UniqueName: \"kubernetes.io/projected/70cc1105-fd62-42e9-b6cb-612adea92724-kube-api-access-rp87l\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.418875 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/70cc1105-fd62-42e9-b6cb-612adea92724-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.954851 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" event={"ID":"70cc1105-fd62-42e9-b6cb-612adea92724","Type":"ContainerDied","Data":"6c58c661af0e375467a4809e919e99a9232be284879feafb6a0c88946e5c128a"} Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.955108 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c58c661af0e375467a4809e919e99a9232be284879feafb6a0c88946e5c128a" Mar 09 09:40:33 crc kubenswrapper[4692]: I0309 09:40:33.954904 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-knbvh" Mar 09 09:40:39 crc kubenswrapper[4692]: I0309 09:40:39.549799 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:39 crc kubenswrapper[4692]: I0309 09:40:39.556763 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"swift-storage-0\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:39 crc kubenswrapper[4692]: I0309 09:40:39.607786 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:40:40 crc kubenswrapper[4692]: I0309 09:40:40.043102 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:40:40 crc kubenswrapper[4692]: W0309 09:40:40.050419 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc01b6df5_bc9c_4abd_9ce5_e928b06580c7.slice/crio-77a6f89143fe945e1692feadac74c09e67eb12d755f566307c8fb1e988a0c098 WatchSource:0}: Error finding container 77a6f89143fe945e1692feadac74c09e67eb12d755f566307c8fb1e988a0c098: Status 404 returned error can't find the container with id 77a6f89143fe945e1692feadac74c09e67eb12d755f566307c8fb1e988a0c098 Mar 09 09:40:41 crc kubenswrapper[4692]: I0309 09:40:41.018755 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"2ea0a775295e7a2ebd710a249b8eef9e502086a8adca32469d0a7e5f0360946b"} Mar 09 09:40:41 crc kubenswrapper[4692]: I0309 09:40:41.018976 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"3c66beccc4342856159e6f72d1d1ae309389f4f7b64191912f7f8ea930ad529f"} Mar 09 09:40:41 crc kubenswrapper[4692]: I0309 09:40:41.018986 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"c280851b4e3b0f461e368eb3548350b7ce5623cbea4928739b6a1c9c9b6e4b82"} Mar 09 09:40:41 crc kubenswrapper[4692]: I0309 09:40:41.018996 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"bb413cb17c31ff233ceedc8a35e86e5b01a697b0c889b4a43599446ef0c3e78d"} Mar 09 09:40:41 crc kubenswrapper[4692]: I0309 09:40:41.019007 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"c4ecffd2dfe8323c86c3932e7f158bba159aef48ce3a7407fa6ec794dcb21fb9"} Mar 09 09:40:41 crc kubenswrapper[4692]: I0309 09:40:41.019016 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"ab65f21b397d18fb36d0d4050eebb38914e73ef3e17f18703e7a399455036458"} Mar 09 09:40:41 crc kubenswrapper[4692]: I0309 09:40:41.019035 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"77a6f89143fe945e1692feadac74c09e67eb12d755f566307c8fb1e988a0c098"} Mar 09 09:40:42 crc kubenswrapper[4692]: I0309 09:40:42.042787 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"06be2bd359a325cfb2ef0967a7db004ec3b7c9adb30c68c9aa995111081731ee"} Mar 09 09:40:42 crc kubenswrapper[4692]: I0309 09:40:42.043107 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"d96fdc6db8ea6a8d4119dbbc1739cb93789eee6ec1375ebdcdbbcaac8fce23fb"} Mar 09 09:40:42 crc kubenswrapper[4692]: I0309 09:40:42.043120 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"e93f97a75cd47e8689825ca7c8a46036015ce8c783fbd946e4b5aae4c8f79b03"} Mar 09 09:40:42 crc kubenswrapper[4692]: I0309 09:40:42.043130 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"9c329076d4b5930b205901b8cdceb024a0470b51a44c26609e18b30b6e9ff6b9"} Mar 09 09:40:42 crc kubenswrapper[4692]: I0309 09:40:42.043143 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"4b04d6a9a6046a7cf5adeb6b3544acc38445e80b0d6ec6bd670e88aac9bee508"} Mar 09 09:40:43 crc kubenswrapper[4692]: I0309 09:40:43.062563 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"564d437ed65b90cb81f73d2a130ee67b0cedeedb41b6b4d9a24f4f61cfa50bcd"} Mar 09 09:40:43 crc kubenswrapper[4692]: I0309 09:40:43.062889 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"58b3d88d99997185e1fe0c65177b1557bbf0f1600dfcd4c1a78a97dfe4cc3869"} Mar 09 09:40:43 crc kubenswrapper[4692]: I0309 09:40:43.062900 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"ab12c78b3d286b45efb3c2c57c2505919e7441ae58bbc75580420db1252bd058"} Mar 09 09:40:43 crc kubenswrapper[4692]: I0309 09:40:43.062910 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"2394ec11824f008a37542215a330958d91bdf6a7d65a1fd73ce74901131426d5"} Mar 09 09:40:43 crc kubenswrapper[4692]: I0309 09:40:43.062919 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerStarted","Data":"703b6e2cbdc86106d9e4b0e1d0f94ab5534c785c45615232e8488b6c7384526e"} Mar 09 09:40:43 crc kubenswrapper[4692]: I0309 09:40:43.111854 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-0" podStartSLOduration=21.111832751 podStartE2EDuration="21.111832751s" podCreationTimestamp="2026-03-09 09:40:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:40:43.097944756 +0000 UTC m=+1243.922680337" watchObservedRunningTime="2026-03-09 09:40:43.111832751 +0000 UTC m=+1243.936568352" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.732062 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-proxy-bfcdc9759-4998p"] Mar 09 09:40:50 crc kubenswrapper[4692]: E0309 09:40:50.733233 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70cc1105-fd62-42e9-b6cb-612adea92724" containerName="swift-ring-rebalance" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.733272 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="70cc1105-fd62-42e9-b6cb-612adea92724" containerName="swift-ring-rebalance" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.733436 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="70cc1105-fd62-42e9-b6cb-612adea92724" containerName="swift-ring-rebalance" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.734423 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.736504 4692 reflector.go:368] Caches populated for *v1.Secret from object-"swift-kuttl-tests"/"swift-proxy-config-data" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.791502 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-bfcdc9759-4998p"] Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.848706 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-etc-swift\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.849049 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-log-httpd\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.849286 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-run-httpd\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.849438 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23df0a-a16c-4a17-8308-909043e28377-config-data\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.849561 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7cf6\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-kube-api-access-z7cf6\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.950775 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-etc-swift\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.950857 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-log-httpd\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.950903 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-run-httpd\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.950946 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23df0a-a16c-4a17-8308-909043e28377-config-data\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.950990 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7cf6\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-kube-api-access-z7cf6\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.951632 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-run-httpd\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.952124 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-log-httpd\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.958885 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23df0a-a16c-4a17-8308-909043e28377-config-data\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.968345 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-etc-swift\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:50 crc kubenswrapper[4692]: I0309 09:40:50.980962 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7cf6\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-kube-api-access-z7cf6\") pod \"swift-proxy-bfcdc9759-4998p\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:51 crc kubenswrapper[4692]: I0309 09:40:51.095271 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:51 crc kubenswrapper[4692]: I0309 09:40:51.514541 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-bfcdc9759-4998p"] Mar 09 09:40:51 crc kubenswrapper[4692]: W0309 09:40:51.518554 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd23df0a_a16c_4a17_8308_909043e28377.slice/crio-e6766a81cb2b796bea4981966dc1d166f9eebd537b4e8e93480e92baf4238e1a WatchSource:0}: Error finding container e6766a81cb2b796bea4981966dc1d166f9eebd537b4e8e93480e92baf4238e1a: Status 404 returned error can't find the container with id e6766a81cb2b796bea4981966dc1d166f9eebd537b4e8e93480e92baf4238e1a Mar 09 09:40:52 crc kubenswrapper[4692]: I0309 09:40:52.127554 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" event={"ID":"bd23df0a-a16c-4a17-8308-909043e28377","Type":"ContainerStarted","Data":"95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83"} Mar 09 09:40:52 crc kubenswrapper[4692]: I0309 09:40:52.127866 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" event={"ID":"bd23df0a-a16c-4a17-8308-909043e28377","Type":"ContainerStarted","Data":"bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9"} Mar 09 09:40:52 crc kubenswrapper[4692]: I0309 09:40:52.127882 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:52 crc kubenswrapper[4692]: I0309 09:40:52.127891 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" event={"ID":"bd23df0a-a16c-4a17-8308-909043e28377","Type":"ContainerStarted","Data":"e6766a81cb2b796bea4981966dc1d166f9eebd537b4e8e93480e92baf4238e1a"} Mar 09 09:40:52 crc kubenswrapper[4692]: I0309 09:40:52.145188 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" podStartSLOduration=2.145155548 podStartE2EDuration="2.145155548s" podCreationTimestamp="2026-03-09 09:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:40:52.141117193 +0000 UTC m=+1252.965852774" watchObservedRunningTime="2026-03-09 09:40:52.145155548 +0000 UTC m=+1252.969891129" Mar 09 09:40:53 crc kubenswrapper[4692]: I0309 09:40:53.134480 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:56 crc kubenswrapper[4692]: I0309 09:40:56.100823 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:56 crc kubenswrapper[4692]: I0309 09:40:56.102226 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:40:57 crc kubenswrapper[4692]: I0309 09:40:57.971573 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw"] Mar 09 09:40:57 crc kubenswrapper[4692]: I0309 09:40:57.972788 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:57 crc kubenswrapper[4692]: I0309 09:40:57.975869 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:40:57 crc kubenswrapper[4692]: I0309 09:40:57.978900 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:40:57 crc kubenswrapper[4692]: I0309 09:40:57.982245 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw"] Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.047316 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsdvd\" (UniqueName: \"kubernetes.io/projected/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-kube-api-access-wsdvd\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.047385 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-ring-data-devices\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.047422 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-scripts\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.047491 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-etc-swift\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.047528 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-dispersionconf\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.047551 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-swiftconf\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.148754 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-etc-swift\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.148846 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-dispersionconf\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.148882 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-swiftconf\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.148921 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsdvd\" (UniqueName: \"kubernetes.io/projected/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-kube-api-access-wsdvd\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.148958 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-ring-data-devices\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.148981 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-scripts\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.149326 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-etc-swift\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.149710 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-scripts\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.149924 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-ring-data-devices\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.154461 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-dispersionconf\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.154625 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-swiftconf\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.165768 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsdvd\" (UniqueName: \"kubernetes.io/projected/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-kube-api-access-wsdvd\") pod \"swift-ring-rebalance-debug-ssmkw\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.302864 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:40:58 crc kubenswrapper[4692]: I0309 09:40:58.710617 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw"] Mar 09 09:40:59 crc kubenswrapper[4692]: I0309 09:40:59.178280 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" event={"ID":"67d856c8-ce30-4ba9-93bf-1720ff29a0cd","Type":"ContainerStarted","Data":"74e1a884f7f9f8904a07b1b8011e4ba8e3c1816f8dcff5525e38a3d54150a0b7"} Mar 09 09:40:59 crc kubenswrapper[4692]: I0309 09:40:59.179298 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" event={"ID":"67d856c8-ce30-4ba9-93bf-1720ff29a0cd","Type":"ContainerStarted","Data":"0ecd1cff4d9e3a61afd20f2713f7a2d63cf48d95de65f3e357968cdb0e625af7"} Mar 09 09:40:59 crc kubenswrapper[4692]: I0309 09:40:59.201439 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" podStartSLOduration=2.201423383 podStartE2EDuration="2.201423383s" podCreationTimestamp="2026-03-09 09:40:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:40:59.197545173 +0000 UTC m=+1260.022280754" watchObservedRunningTime="2026-03-09 09:40:59.201423383 +0000 UTC m=+1260.026158964" Mar 09 09:41:01 crc kubenswrapper[4692]: I0309 09:41:01.193485 4692 generic.go:334] "Generic (PLEG): container finished" podID="67d856c8-ce30-4ba9-93bf-1720ff29a0cd" containerID="74e1a884f7f9f8904a07b1b8011e4ba8e3c1816f8dcff5525e38a3d54150a0b7" exitCode=0 Mar 09 09:41:01 crc kubenswrapper[4692]: I0309 09:41:01.193530 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" event={"ID":"67d856c8-ce30-4ba9-93bf-1720ff29a0cd","Type":"ContainerDied","Data":"74e1a884f7f9f8904a07b1b8011e4ba8e3c1816f8dcff5525e38a3d54150a0b7"} Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.519421 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.545019 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw"] Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.553405 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw"] Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.618236 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-swiftconf\") pod \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.618298 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-ring-data-devices\") pod \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.618378 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-scripts\") pod \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.618423 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsdvd\" (UniqueName: \"kubernetes.io/projected/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-kube-api-access-wsdvd\") pod \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.618470 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-dispersionconf\") pod \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.618500 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-etc-swift\") pod \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\" (UID: \"67d856c8-ce30-4ba9-93bf-1720ff29a0cd\") " Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.619027 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "67d856c8-ce30-4ba9-93bf-1720ff29a0cd" (UID: "67d856c8-ce30-4ba9-93bf-1720ff29a0cd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.619406 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "67d856c8-ce30-4ba9-93bf-1720ff29a0cd" (UID: "67d856c8-ce30-4ba9-93bf-1720ff29a0cd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.623347 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-kube-api-access-wsdvd" (OuterVolumeSpecName: "kube-api-access-wsdvd") pod "67d856c8-ce30-4ba9-93bf-1720ff29a0cd" (UID: "67d856c8-ce30-4ba9-93bf-1720ff29a0cd"). InnerVolumeSpecName "kube-api-access-wsdvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.630566 4692 scope.go:117] "RemoveContainer" containerID="c47374ee0cef70824c305d4e806b22e9b170ce84fa698597c397a090752e0f3e" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.638736 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "67d856c8-ce30-4ba9-93bf-1720ff29a0cd" (UID: "67d856c8-ce30-4ba9-93bf-1720ff29a0cd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.639814 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-scripts" (OuterVolumeSpecName: "scripts") pod "67d856c8-ce30-4ba9-93bf-1720ff29a0cd" (UID: "67d856c8-ce30-4ba9-93bf-1720ff29a0cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.648812 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "67d856c8-ce30-4ba9-93bf-1720ff29a0cd" (UID: "67d856c8-ce30-4ba9-93bf-1720ff29a0cd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.720863 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.720898 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsdvd\" (UniqueName: \"kubernetes.io/projected/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-kube-api-access-wsdvd\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.720911 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.720922 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.720933 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.720944 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/67d856c8-ce30-4ba9-93bf-1720ff29a0cd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.743723 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm"] Mar 09 09:41:02 crc kubenswrapper[4692]: E0309 09:41:02.744110 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d856c8-ce30-4ba9-93bf-1720ff29a0cd" containerName="swift-ring-rebalance" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.744127 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d856c8-ce30-4ba9-93bf-1720ff29a0cd" containerName="swift-ring-rebalance" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.744273 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d856c8-ce30-4ba9-93bf-1720ff29a0cd" containerName="swift-ring-rebalance" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.744732 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.754969 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm"] Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.822094 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-scripts\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.822204 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-swiftconf\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.822260 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlz8g\" (UniqueName: \"kubernetes.io/projected/9c09be70-7620-415d-be2a-8b655a0070d2-kube-api-access-xlz8g\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.822287 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-ring-data-devices\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.822408 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9c09be70-7620-415d-be2a-8b655a0070d2-etc-swift\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.822525 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-dispersionconf\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.923913 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlz8g\" (UniqueName: \"kubernetes.io/projected/9c09be70-7620-415d-be2a-8b655a0070d2-kube-api-access-xlz8g\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.923994 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-ring-data-devices\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.924026 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9c09be70-7620-415d-be2a-8b655a0070d2-etc-swift\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.924730 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9c09be70-7620-415d-be2a-8b655a0070d2-etc-swift\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.924912 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-ring-data-devices\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.924976 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-dispersionconf\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.925078 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-scripts\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.925523 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-swiftconf\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.925638 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-scripts\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.928480 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-dispersionconf\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.929581 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-swiftconf\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:02 crc kubenswrapper[4692]: I0309 09:41:02.939378 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlz8g\" (UniqueName: \"kubernetes.io/projected/9c09be70-7620-415d-be2a-8b655a0070d2-kube-api-access-xlz8g\") pod \"swift-ring-rebalance-debug-zzmrm\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:03 crc kubenswrapper[4692]: I0309 09:41:03.064893 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:03 crc kubenswrapper[4692]: I0309 09:41:03.229785 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ecd1cff4d9e3a61afd20f2713f7a2d63cf48d95de65f3e357968cdb0e625af7" Mar 09 09:41:03 crc kubenswrapper[4692]: I0309 09:41:03.230125 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ssmkw" Mar 09 09:41:03 crc kubenswrapper[4692]: I0309 09:41:03.562438 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm"] Mar 09 09:41:03 crc kubenswrapper[4692]: W0309 09:41:03.566924 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c09be70_7620_415d_be2a_8b655a0070d2.slice/crio-53fdce6aa1f78eeeb877be318d9988a689ef2b50d04ca38b996e6d0cb064e5cd WatchSource:0}: Error finding container 53fdce6aa1f78eeeb877be318d9988a689ef2b50d04ca38b996e6d0cb064e5cd: Status 404 returned error can't find the container with id 53fdce6aa1f78eeeb877be318d9988a689ef2b50d04ca38b996e6d0cb064e5cd Mar 09 09:41:04 crc kubenswrapper[4692]: I0309 09:41:04.082568 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d856c8-ce30-4ba9-93bf-1720ff29a0cd" path="/var/lib/kubelet/pods/67d856c8-ce30-4ba9-93bf-1720ff29a0cd/volumes" Mar 09 09:41:04 crc kubenswrapper[4692]: I0309 09:41:04.238367 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" event={"ID":"9c09be70-7620-415d-be2a-8b655a0070d2","Type":"ContainerStarted","Data":"138283a861b96ca4ff869cad6bc6e67d44a003b52bddd47c62434dce3f45ebd5"} Mar 09 09:41:04 crc kubenswrapper[4692]: I0309 09:41:04.238410 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" event={"ID":"9c09be70-7620-415d-be2a-8b655a0070d2","Type":"ContainerStarted","Data":"53fdce6aa1f78eeeb877be318d9988a689ef2b50d04ca38b996e6d0cb064e5cd"} Mar 09 09:41:04 crc kubenswrapper[4692]: I0309 09:41:04.262949 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" podStartSLOduration=2.262928696 podStartE2EDuration="2.262928696s" podCreationTimestamp="2026-03-09 09:41:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:41:04.259046016 +0000 UTC m=+1265.083781607" watchObservedRunningTime="2026-03-09 09:41:04.262928696 +0000 UTC m=+1265.087664277" Mar 09 09:41:05 crc kubenswrapper[4692]: I0309 09:41:05.246083 4692 generic.go:334] "Generic (PLEG): container finished" podID="9c09be70-7620-415d-be2a-8b655a0070d2" containerID="138283a861b96ca4ff869cad6bc6e67d44a003b52bddd47c62434dce3f45ebd5" exitCode=0 Mar 09 09:41:05 crc kubenswrapper[4692]: I0309 09:41:05.246146 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" event={"ID":"9c09be70-7620-415d-be2a-8b655a0070d2","Type":"ContainerDied","Data":"138283a861b96ca4ff869cad6bc6e67d44a003b52bddd47c62434dce3f45ebd5"} Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.521143 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.550054 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm"] Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.554964 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm"] Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.588561 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-swiftconf\") pod \"9c09be70-7620-415d-be2a-8b655a0070d2\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.588628 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlz8g\" (UniqueName: \"kubernetes.io/projected/9c09be70-7620-415d-be2a-8b655a0070d2-kube-api-access-xlz8g\") pod \"9c09be70-7620-415d-be2a-8b655a0070d2\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.588649 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9c09be70-7620-415d-be2a-8b655a0070d2-etc-swift\") pod \"9c09be70-7620-415d-be2a-8b655a0070d2\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.588672 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-ring-data-devices\") pod \"9c09be70-7620-415d-be2a-8b655a0070d2\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.588758 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-dispersionconf\") pod \"9c09be70-7620-415d-be2a-8b655a0070d2\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.588817 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-scripts\") pod \"9c09be70-7620-415d-be2a-8b655a0070d2\" (UID: \"9c09be70-7620-415d-be2a-8b655a0070d2\") " Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.590333 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9c09be70-7620-415d-be2a-8b655a0070d2" (UID: "9c09be70-7620-415d-be2a-8b655a0070d2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.590699 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c09be70-7620-415d-be2a-8b655a0070d2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9c09be70-7620-415d-be2a-8b655a0070d2" (UID: "9c09be70-7620-415d-be2a-8b655a0070d2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.594092 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c09be70-7620-415d-be2a-8b655a0070d2-kube-api-access-xlz8g" (OuterVolumeSpecName: "kube-api-access-xlz8g") pod "9c09be70-7620-415d-be2a-8b655a0070d2" (UID: "9c09be70-7620-415d-be2a-8b655a0070d2"). InnerVolumeSpecName "kube-api-access-xlz8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.609103 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9c09be70-7620-415d-be2a-8b655a0070d2" (UID: "9c09be70-7620-415d-be2a-8b655a0070d2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.609470 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9c09be70-7620-415d-be2a-8b655a0070d2" (UID: "9c09be70-7620-415d-be2a-8b655a0070d2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.611573 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-scripts" (OuterVolumeSpecName: "scripts") pod "9c09be70-7620-415d-be2a-8b655a0070d2" (UID: "9c09be70-7620-415d-be2a-8b655a0070d2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.691196 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.691247 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlz8g\" (UniqueName: \"kubernetes.io/projected/9c09be70-7620-415d-be2a-8b655a0070d2-kube-api-access-xlz8g\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.691261 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9c09be70-7620-415d-be2a-8b655a0070d2-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.691273 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.691287 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9c09be70-7620-415d-be2a-8b655a0070d2-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:06 crc kubenswrapper[4692]: I0309 09:41:06.691300 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9c09be70-7620-415d-be2a-8b655a0070d2-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:41:07 crc kubenswrapper[4692]: I0309 09:41:07.261464 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53fdce6aa1f78eeeb877be318d9988a689ef2b50d04ca38b996e6d0cb064e5cd" Mar 09 09:41:07 crc kubenswrapper[4692]: I0309 09:41:07.261515 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zzmrm" Mar 09 09:41:08 crc kubenswrapper[4692]: I0309 09:41:08.086388 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c09be70-7620-415d-be2a-8b655a0070d2" path="/var/lib/kubelet/pods/9c09be70-7620-415d-be2a-8b655a0070d2/volumes" Mar 09 09:41:47 crc kubenswrapper[4692]: I0309 09:41:47.623553 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:41:47 crc kubenswrapper[4692]: I0309 09:41:47.624107 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.140885 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550822-g4wkt"] Mar 09 09:42:00 crc kubenswrapper[4692]: E0309 09:42:00.142231 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c09be70-7620-415d-be2a-8b655a0070d2" containerName="swift-ring-rebalance" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.142248 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c09be70-7620-415d-be2a-8b655a0070d2" containerName="swift-ring-rebalance" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.142555 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c09be70-7620-415d-be2a-8b655a0070d2" containerName="swift-ring-rebalance" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.143254 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.148081 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.148200 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.148104 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.157151 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550822-g4wkt"] Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.160019 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srp8d\" (UniqueName: \"kubernetes.io/projected/059e481e-f65b-4413-969e-edf4666c4e63-kube-api-access-srp8d\") pod \"auto-csr-approver-29550822-g4wkt\" (UID: \"059e481e-f65b-4413-969e-edf4666c4e63\") " pod="openshift-infra/auto-csr-approver-29550822-g4wkt" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.261234 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srp8d\" (UniqueName: \"kubernetes.io/projected/059e481e-f65b-4413-969e-edf4666c4e63-kube-api-access-srp8d\") pod \"auto-csr-approver-29550822-g4wkt\" (UID: \"059e481e-f65b-4413-969e-edf4666c4e63\") " pod="openshift-infra/auto-csr-approver-29550822-g4wkt" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.280891 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srp8d\" (UniqueName: \"kubernetes.io/projected/059e481e-f65b-4413-969e-edf4666c4e63-kube-api-access-srp8d\") pod \"auto-csr-approver-29550822-g4wkt\" (UID: \"059e481e-f65b-4413-969e-edf4666c4e63\") " pod="openshift-infra/auto-csr-approver-29550822-g4wkt" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.477226 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" Mar 09 09:42:00 crc kubenswrapper[4692]: I0309 09:42:00.890159 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550822-g4wkt"] Mar 09 09:42:01 crc kubenswrapper[4692]: I0309 09:42:01.029385 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" event={"ID":"059e481e-f65b-4413-969e-edf4666c4e63","Type":"ContainerStarted","Data":"6b3d4964984870135fcab4f4dc697609890af227f3a6dbae5dfacba5eb0afe93"} Mar 09 09:42:02 crc kubenswrapper[4692]: I0309 09:42:02.041436 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" event={"ID":"059e481e-f65b-4413-969e-edf4666c4e63","Type":"ContainerStarted","Data":"200d9b7a50a0ab84887facd9cd4431b8973fc90227411f1b5e8252dbc9ebdc2c"} Mar 09 09:42:02 crc kubenswrapper[4692]: I0309 09:42:02.060741 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" podStartSLOduration=1.228826012 podStartE2EDuration="2.060716053s" podCreationTimestamp="2026-03-09 09:42:00 +0000 UTC" firstStartedPulling="2026-03-09 09:42:00.895692843 +0000 UTC m=+1321.720428424" lastFinishedPulling="2026-03-09 09:42:01.727582884 +0000 UTC m=+1322.552318465" observedRunningTime="2026-03-09 09:42:02.055412793 +0000 UTC m=+1322.880148374" watchObservedRunningTime="2026-03-09 09:42:02.060716053 +0000 UTC m=+1322.885451634" Mar 09 09:42:03 crc kubenswrapper[4692]: I0309 09:42:03.050795 4692 generic.go:334] "Generic (PLEG): container finished" podID="059e481e-f65b-4413-969e-edf4666c4e63" containerID="200d9b7a50a0ab84887facd9cd4431b8973fc90227411f1b5e8252dbc9ebdc2c" exitCode=0 Mar 09 09:42:03 crc kubenswrapper[4692]: I0309 09:42:03.051015 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" event={"ID":"059e481e-f65b-4413-969e-edf4666c4e63","Type":"ContainerDied","Data":"200d9b7a50a0ab84887facd9cd4431b8973fc90227411f1b5e8252dbc9ebdc2c"} Mar 09 09:42:04 crc kubenswrapper[4692]: I0309 09:42:04.322838 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" Mar 09 09:42:04 crc kubenswrapper[4692]: I0309 09:42:04.518150 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srp8d\" (UniqueName: \"kubernetes.io/projected/059e481e-f65b-4413-969e-edf4666c4e63-kube-api-access-srp8d\") pod \"059e481e-f65b-4413-969e-edf4666c4e63\" (UID: \"059e481e-f65b-4413-969e-edf4666c4e63\") " Mar 09 09:42:04 crc kubenswrapper[4692]: I0309 09:42:04.524392 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/059e481e-f65b-4413-969e-edf4666c4e63-kube-api-access-srp8d" (OuterVolumeSpecName: "kube-api-access-srp8d") pod "059e481e-f65b-4413-969e-edf4666c4e63" (UID: "059e481e-f65b-4413-969e-edf4666c4e63"). InnerVolumeSpecName "kube-api-access-srp8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:42:04 crc kubenswrapper[4692]: I0309 09:42:04.620472 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srp8d\" (UniqueName: \"kubernetes.io/projected/059e481e-f65b-4413-969e-edf4666c4e63-kube-api-access-srp8d\") on node \"crc\" DevicePath \"\"" Mar 09 09:42:05 crc kubenswrapper[4692]: I0309 09:42:05.067777 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" event={"ID":"059e481e-f65b-4413-969e-edf4666c4e63","Type":"ContainerDied","Data":"6b3d4964984870135fcab4f4dc697609890af227f3a6dbae5dfacba5eb0afe93"} Mar 09 09:42:05 crc kubenswrapper[4692]: I0309 09:42:05.067820 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b3d4964984870135fcab4f4dc697609890af227f3a6dbae5dfacba5eb0afe93" Mar 09 09:42:05 crc kubenswrapper[4692]: I0309 09:42:05.067830 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550822-g4wkt" Mar 09 09:42:05 crc kubenswrapper[4692]: I0309 09:42:05.112099 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550816-j2x6f"] Mar 09 09:42:05 crc kubenswrapper[4692]: I0309 09:42:05.118714 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550816-j2x6f"] Mar 09 09:42:06 crc kubenswrapper[4692]: I0309 09:42:06.080368 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="355a2d83-082d-4181-a67e-01373c53a52a" path="/var/lib/kubelet/pods/355a2d83-082d-4181-a67e-01373c53a52a/volumes" Mar 09 09:42:17 crc kubenswrapper[4692]: I0309 09:42:17.624470 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:42:17 crc kubenswrapper[4692]: I0309 09:42:17.624928 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.668982 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jvwq2"] Mar 09 09:42:42 crc kubenswrapper[4692]: E0309 09:42:42.669996 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="059e481e-f65b-4413-969e-edf4666c4e63" containerName="oc" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.670010 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="059e481e-f65b-4413-969e-edf4666c4e63" containerName="oc" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.670136 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="059e481e-f65b-4413-969e-edf4666c4e63" containerName="oc" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.671122 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.694639 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvwq2"] Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.743084 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-catalog-content\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.743187 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr9sw\" (UniqueName: \"kubernetes.io/projected/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-kube-api-access-cr9sw\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.743308 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-utilities\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.844273 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-utilities\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.844399 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-catalog-content\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.844803 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-utilities\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.844852 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-catalog-content\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.844937 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr9sw\" (UniqueName: \"kubernetes.io/projected/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-kube-api-access-cr9sw\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.876432 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr9sw\" (UniqueName: \"kubernetes.io/projected/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-kube-api-access-cr9sw\") pod \"redhat-operators-jvwq2\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:42 crc kubenswrapper[4692]: I0309 09:42:42.990930 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:43 crc kubenswrapper[4692]: I0309 09:42:43.468759 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvwq2"] Mar 09 09:42:43 crc kubenswrapper[4692]: W0309 09:42:43.476004 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf964e8d8_955b_41dc_a1c4_7d8bcb25c235.slice/crio-bf564d4c34df7e7a6838ae283d97c89862b1727696491deb48cac030397fa602 WatchSource:0}: Error finding container bf564d4c34df7e7a6838ae283d97c89862b1727696491deb48cac030397fa602: Status 404 returned error can't find the container with id bf564d4c34df7e7a6838ae283d97c89862b1727696491deb48cac030397fa602 Mar 09 09:42:44 crc kubenswrapper[4692]: I0309 09:42:44.337153 4692 generic.go:334] "Generic (PLEG): container finished" podID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerID="27736a07381b76cfe99092e19389a89a59fe25c6d6bd0d2edba46a3f2c25468a" exitCode=0 Mar 09 09:42:44 crc kubenswrapper[4692]: I0309 09:42:44.337368 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwq2" event={"ID":"f964e8d8-955b-41dc-a1c4-7d8bcb25c235","Type":"ContainerDied","Data":"27736a07381b76cfe99092e19389a89a59fe25c6d6bd0d2edba46a3f2c25468a"} Mar 09 09:42:44 crc kubenswrapper[4692]: I0309 09:42:44.337465 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwq2" event={"ID":"f964e8d8-955b-41dc-a1c4-7d8bcb25c235","Type":"ContainerStarted","Data":"bf564d4c34df7e7a6838ae283d97c89862b1727696491deb48cac030397fa602"} Mar 09 09:42:46 crc kubenswrapper[4692]: I0309 09:42:46.353245 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwq2" event={"ID":"f964e8d8-955b-41dc-a1c4-7d8bcb25c235","Type":"ContainerStarted","Data":"bb290b81594a47a7f8f8efea8ea9c7995c14c793c207ea8ba9a5b5358b4cb528"} Mar 09 09:42:47 crc kubenswrapper[4692]: I0309 09:42:47.361759 4692 generic.go:334] "Generic (PLEG): container finished" podID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerID="bb290b81594a47a7f8f8efea8ea9c7995c14c793c207ea8ba9a5b5358b4cb528" exitCode=0 Mar 09 09:42:47 crc kubenswrapper[4692]: I0309 09:42:47.361865 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwq2" event={"ID":"f964e8d8-955b-41dc-a1c4-7d8bcb25c235","Type":"ContainerDied","Data":"bb290b81594a47a7f8f8efea8ea9c7995c14c793c207ea8ba9a5b5358b4cb528"} Mar 09 09:42:47 crc kubenswrapper[4692]: I0309 09:42:47.624534 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:42:47 crc kubenswrapper[4692]: I0309 09:42:47.624599 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:42:47 crc kubenswrapper[4692]: I0309 09:42:47.624646 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:42:47 crc kubenswrapper[4692]: I0309 09:42:47.625276 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc5528f2b79383c92cd6d5389f75e574b39b690b0caddba49d4c6810fb3c85bb"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:42:47 crc kubenswrapper[4692]: I0309 09:42:47.625349 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://dc5528f2b79383c92cd6d5389f75e574b39b690b0caddba49d4c6810fb3c85bb" gracePeriod=600 Mar 09 09:42:48 crc kubenswrapper[4692]: I0309 09:42:48.373243 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="dc5528f2b79383c92cd6d5389f75e574b39b690b0caddba49d4c6810fb3c85bb" exitCode=0 Mar 09 09:42:48 crc kubenswrapper[4692]: I0309 09:42:48.373327 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"dc5528f2b79383c92cd6d5389f75e574b39b690b0caddba49d4c6810fb3c85bb"} Mar 09 09:42:48 crc kubenswrapper[4692]: I0309 09:42:48.373714 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46"} Mar 09 09:42:48 crc kubenswrapper[4692]: I0309 09:42:48.373740 4692 scope.go:117] "RemoveContainer" containerID="a00e663af9bd3bc83fcf84afc788c06c37b9025a209fe4f4ecbe66f9658cebf6" Mar 09 09:42:49 crc kubenswrapper[4692]: I0309 09:42:49.386920 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwq2" event={"ID":"f964e8d8-955b-41dc-a1c4-7d8bcb25c235","Type":"ContainerStarted","Data":"78ff6695ffccc613d5fca07ae16e93c8f26fa4c35cc450edcf136eb0b1432292"} Mar 09 09:42:49 crc kubenswrapper[4692]: I0309 09:42:49.410748 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jvwq2" podStartSLOduration=2.985837891 podStartE2EDuration="7.41072673s" podCreationTimestamp="2026-03-09 09:42:42 +0000 UTC" firstStartedPulling="2026-03-09 09:42:44.338828707 +0000 UTC m=+1365.163564278" lastFinishedPulling="2026-03-09 09:42:48.763717536 +0000 UTC m=+1369.588453117" observedRunningTime="2026-03-09 09:42:49.402647191 +0000 UTC m=+1370.227382772" watchObservedRunningTime="2026-03-09 09:42:49.41072673 +0000 UTC m=+1370.235462311" Mar 09 09:42:52 crc kubenswrapper[4692]: I0309 09:42:52.991189 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:52 crc kubenswrapper[4692]: I0309 09:42:52.991648 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:42:54 crc kubenswrapper[4692]: I0309 09:42:54.029609 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jvwq2" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="registry-server" probeResult="failure" output=< Mar 09 09:42:54 crc kubenswrapper[4692]: timeout: failed to connect service ":50051" within 1s Mar 09 09:42:54 crc kubenswrapper[4692]: > Mar 09 09:43:02 crc kubenswrapper[4692]: I0309 09:43:02.784775 4692 scope.go:117] "RemoveContainer" containerID="2428897e623dc39d6402064f77ddc16fe944452aeac8b1afdc908e25ff32a7e6" Mar 09 09:43:03 crc kubenswrapper[4692]: I0309 09:43:03.037140 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:43:03 crc kubenswrapper[4692]: I0309 09:43:03.088567 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:43:06 crc kubenswrapper[4692]: I0309 09:43:06.099030 4692 prober.go:107] "Probe failed" probeType="Readiness" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 09 09:43:06 crc kubenswrapper[4692]: I0309 09:43:06.655188 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvwq2"] Mar 09 09:43:06 crc kubenswrapper[4692]: I0309 09:43:06.655419 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jvwq2" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="registry-server" containerID="cri-o://78ff6695ffccc613d5fca07ae16e93c8f26fa4c35cc450edcf136eb0b1432292" gracePeriod=2 Mar 09 09:43:06 crc kubenswrapper[4692]: I0309 09:43:06.886313 4692 generic.go:334] "Generic (PLEG): container finished" podID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerID="78ff6695ffccc613d5fca07ae16e93c8f26fa4c35cc450edcf136eb0b1432292" exitCode=0 Mar 09 09:43:06 crc kubenswrapper[4692]: I0309 09:43:06.886359 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwq2" event={"ID":"f964e8d8-955b-41dc-a1c4-7d8bcb25c235","Type":"ContainerDied","Data":"78ff6695ffccc613d5fca07ae16e93c8f26fa4c35cc450edcf136eb0b1432292"} Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.046542 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.211730 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-catalog-content\") pod \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.211839 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr9sw\" (UniqueName: \"kubernetes.io/projected/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-kube-api-access-cr9sw\") pod \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.211884 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-utilities\") pod \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\" (UID: \"f964e8d8-955b-41dc-a1c4-7d8bcb25c235\") " Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.214231 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-utilities" (OuterVolumeSpecName: "utilities") pod "f964e8d8-955b-41dc-a1c4-7d8bcb25c235" (UID: "f964e8d8-955b-41dc-a1c4-7d8bcb25c235"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.220344 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-kube-api-access-cr9sw" (OuterVolumeSpecName: "kube-api-access-cr9sw") pod "f964e8d8-955b-41dc-a1c4-7d8bcb25c235" (UID: "f964e8d8-955b-41dc-a1c4-7d8bcb25c235"). InnerVolumeSpecName "kube-api-access-cr9sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.313231 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr9sw\" (UniqueName: \"kubernetes.io/projected/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-kube-api-access-cr9sw\") on node \"crc\" DevicePath \"\"" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.313557 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.336282 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f964e8d8-955b-41dc-a1c4-7d8bcb25c235" (UID: "f964e8d8-955b-41dc-a1c4-7d8bcb25c235"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.414790 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f964e8d8-955b-41dc-a1c4-7d8bcb25c235-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.895984 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwq2" event={"ID":"f964e8d8-955b-41dc-a1c4-7d8bcb25c235","Type":"ContainerDied","Data":"bf564d4c34df7e7a6838ae283d97c89862b1727696491deb48cac030397fa602"} Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.896039 4692 scope.go:117] "RemoveContainer" containerID="78ff6695ffccc613d5fca07ae16e93c8f26fa4c35cc450edcf136eb0b1432292" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.896068 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwq2" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.921460 4692 scope.go:117] "RemoveContainer" containerID="bb290b81594a47a7f8f8efea8ea9c7995c14c793c207ea8ba9a5b5358b4cb528" Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.936056 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvwq2"] Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.944256 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jvwq2"] Mar 09 09:43:07 crc kubenswrapper[4692]: I0309 09:43:07.947727 4692 scope.go:117] "RemoveContainer" containerID="27736a07381b76cfe99092e19389a89a59fe25c6d6bd0d2edba46a3f2c25468a" Mar 09 09:43:08 crc kubenswrapper[4692]: I0309 09:43:08.080987 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" path="/var/lib/kubelet/pods/f964e8d8-955b-41dc-a1c4-7d8bcb25c235/volumes" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.146400 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550824-cjjf5"] Mar 09 09:44:00 crc kubenswrapper[4692]: E0309 09:44:00.147134 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="extract-utilities" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.147146 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="extract-utilities" Mar 09 09:44:00 crc kubenswrapper[4692]: E0309 09:44:00.147180 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="extract-content" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.147186 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="extract-content" Mar 09 09:44:00 crc kubenswrapper[4692]: E0309 09:44:00.147209 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="registry-server" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.147216 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="registry-server" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.147335 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f964e8d8-955b-41dc-a1c4-7d8bcb25c235" containerName="registry-server" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.147763 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550824-cjjf5" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.149528 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.149594 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.149741 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.158410 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550824-cjjf5"] Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.284547 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcvrq\" (UniqueName: \"kubernetes.io/projected/ac8b5345-e2ae-419d-8de5-b36244b0cf42-kube-api-access-xcvrq\") pod \"auto-csr-approver-29550824-cjjf5\" (UID: \"ac8b5345-e2ae-419d-8de5-b36244b0cf42\") " pod="openshift-infra/auto-csr-approver-29550824-cjjf5" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.386299 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcvrq\" (UniqueName: \"kubernetes.io/projected/ac8b5345-e2ae-419d-8de5-b36244b0cf42-kube-api-access-xcvrq\") pod \"auto-csr-approver-29550824-cjjf5\" (UID: \"ac8b5345-e2ae-419d-8de5-b36244b0cf42\") " pod="openshift-infra/auto-csr-approver-29550824-cjjf5" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.404874 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcvrq\" (UniqueName: \"kubernetes.io/projected/ac8b5345-e2ae-419d-8de5-b36244b0cf42-kube-api-access-xcvrq\") pod \"auto-csr-approver-29550824-cjjf5\" (UID: \"ac8b5345-e2ae-419d-8de5-b36244b0cf42\") " pod="openshift-infra/auto-csr-approver-29550824-cjjf5" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.472799 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550824-cjjf5" Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.889487 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550824-cjjf5"] Mar 09 09:44:00 crc kubenswrapper[4692]: I0309 09:44:00.892734 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 09:44:01 crc kubenswrapper[4692]: I0309 09:44:01.304829 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550824-cjjf5" event={"ID":"ac8b5345-e2ae-419d-8de5-b36244b0cf42","Type":"ContainerStarted","Data":"b1dbb38bf4c2571cff38d46054512eb2f8df1b2c628cdaf0498840d1fa937230"} Mar 09 09:44:02 crc kubenswrapper[4692]: I0309 09:44:02.312536 4692 generic.go:334] "Generic (PLEG): container finished" podID="ac8b5345-e2ae-419d-8de5-b36244b0cf42" containerID="448c9f3423d1ab0e1159af44b656256ded9872b1c632b1fa393e35c24184d419" exitCode=0 Mar 09 09:44:02 crc kubenswrapper[4692]: I0309 09:44:02.312626 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550824-cjjf5" event={"ID":"ac8b5345-e2ae-419d-8de5-b36244b0cf42","Type":"ContainerDied","Data":"448c9f3423d1ab0e1159af44b656256ded9872b1c632b1fa393e35c24184d419"} Mar 09 09:44:03 crc kubenswrapper[4692]: I0309 09:44:03.607804 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550824-cjjf5" Mar 09 09:44:03 crc kubenswrapper[4692]: I0309 09:44:03.728770 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcvrq\" (UniqueName: \"kubernetes.io/projected/ac8b5345-e2ae-419d-8de5-b36244b0cf42-kube-api-access-xcvrq\") pod \"ac8b5345-e2ae-419d-8de5-b36244b0cf42\" (UID: \"ac8b5345-e2ae-419d-8de5-b36244b0cf42\") " Mar 09 09:44:03 crc kubenswrapper[4692]: I0309 09:44:03.733677 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac8b5345-e2ae-419d-8de5-b36244b0cf42-kube-api-access-xcvrq" (OuterVolumeSpecName: "kube-api-access-xcvrq") pod "ac8b5345-e2ae-419d-8de5-b36244b0cf42" (UID: "ac8b5345-e2ae-419d-8de5-b36244b0cf42"). InnerVolumeSpecName "kube-api-access-xcvrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:44:03 crc kubenswrapper[4692]: I0309 09:44:03.830726 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcvrq\" (UniqueName: \"kubernetes.io/projected/ac8b5345-e2ae-419d-8de5-b36244b0cf42-kube-api-access-xcvrq\") on node \"crc\" DevicePath \"\"" Mar 09 09:44:04 crc kubenswrapper[4692]: I0309 09:44:04.329927 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550824-cjjf5" event={"ID":"ac8b5345-e2ae-419d-8de5-b36244b0cf42","Type":"ContainerDied","Data":"b1dbb38bf4c2571cff38d46054512eb2f8df1b2c628cdaf0498840d1fa937230"} Mar 09 09:44:04 crc kubenswrapper[4692]: I0309 09:44:04.329967 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1dbb38bf4c2571cff38d46054512eb2f8df1b2c628cdaf0498840d1fa937230" Mar 09 09:44:04 crc kubenswrapper[4692]: I0309 09:44:04.329999 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550824-cjjf5" Mar 09 09:44:04 crc kubenswrapper[4692]: I0309 09:44:04.670268 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550818-7xlfm"] Mar 09 09:44:04 crc kubenswrapper[4692]: I0309 09:44:04.675681 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550818-7xlfm"] Mar 09 09:44:06 crc kubenswrapper[4692]: I0309 09:44:06.080013 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e5bbd9f-03a6-419e-b495-df023635b82e" path="/var/lib/kubelet/pods/3e5bbd9f-03a6-419e-b495-df023635b82e/volumes" Mar 09 09:44:47 crc kubenswrapper[4692]: I0309 09:44:47.623717 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:44:47 crc kubenswrapper[4692]: I0309 09:44:47.624158 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.760400 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lpv7t"] Mar 09 09:44:56 crc kubenswrapper[4692]: E0309 09:44:56.761237 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac8b5345-e2ae-419d-8de5-b36244b0cf42" containerName="oc" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.761251 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac8b5345-e2ae-419d-8de5-b36244b0cf42" containerName="oc" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.761401 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac8b5345-e2ae-419d-8de5-b36244b0cf42" containerName="oc" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.762542 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.772609 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lpv7t"] Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.817869 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlbqf\" (UniqueName: \"kubernetes.io/projected/96f563ba-b297-4ba0-9747-9a44c127c197-kube-api-access-xlbqf\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.817932 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-utilities\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.817956 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-catalog-content\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.918712 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlbqf\" (UniqueName: \"kubernetes.io/projected/96f563ba-b297-4ba0-9747-9a44c127c197-kube-api-access-xlbqf\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.918765 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-utilities\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.918789 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-catalog-content\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.919242 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-catalog-content\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.919417 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-utilities\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:56 crc kubenswrapper[4692]: I0309 09:44:56.938549 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlbqf\" (UniqueName: \"kubernetes.io/projected/96f563ba-b297-4ba0-9747-9a44c127c197-kube-api-access-xlbqf\") pod \"community-operators-lpv7t\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:57 crc kubenswrapper[4692]: I0309 09:44:57.080584 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:44:57 crc kubenswrapper[4692]: I0309 09:44:57.594764 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lpv7t"] Mar 09 09:44:57 crc kubenswrapper[4692]: I0309 09:44:57.699365 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpv7t" event={"ID":"96f563ba-b297-4ba0-9747-9a44c127c197","Type":"ContainerStarted","Data":"6ed798fda3f1a4e475bb4e00352f3245211d17d5fc2504d41c21963da08feaf7"} Mar 09 09:44:58 crc kubenswrapper[4692]: I0309 09:44:58.709146 4692 generic.go:334] "Generic (PLEG): container finished" podID="96f563ba-b297-4ba0-9747-9a44c127c197" containerID="ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81" exitCode=0 Mar 09 09:44:58 crc kubenswrapper[4692]: I0309 09:44:58.709208 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpv7t" event={"ID":"96f563ba-b297-4ba0-9747-9a44c127c197","Type":"ContainerDied","Data":"ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81"} Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.152461 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf"] Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.154286 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.161238 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.161455 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.166357 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf"] Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.262557 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c316cf6d-8a6f-46fd-a609-9bae80441c51-config-volume\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.262658 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6j2j\" (UniqueName: \"kubernetes.io/projected/c316cf6d-8a6f-46fd-a609-9bae80441c51-kube-api-access-m6j2j\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.262795 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c316cf6d-8a6f-46fd-a609-9bae80441c51-secret-volume\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.364515 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c316cf6d-8a6f-46fd-a609-9bae80441c51-secret-volume\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.364990 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c316cf6d-8a6f-46fd-a609-9bae80441c51-config-volume\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.365051 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6j2j\" (UniqueName: \"kubernetes.io/projected/c316cf6d-8a6f-46fd-a609-9bae80441c51-kube-api-access-m6j2j\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.366823 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c316cf6d-8a6f-46fd-a609-9bae80441c51-config-volume\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.371930 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c316cf6d-8a6f-46fd-a609-9bae80441c51-secret-volume\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.382546 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6j2j\" (UniqueName: \"kubernetes.io/projected/c316cf6d-8a6f-46fd-a609-9bae80441c51-kube-api-access-m6j2j\") pod \"collect-profiles-29550825-7rsqf\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.603536 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.728245 4692 generic.go:334] "Generic (PLEG): container finished" podID="96f563ba-b297-4ba0-9747-9a44c127c197" containerID="d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3" exitCode=0 Mar 09 09:45:00 crc kubenswrapper[4692]: I0309 09:45:00.728285 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpv7t" event={"ID":"96f563ba-b297-4ba0-9747-9a44c127c197","Type":"ContainerDied","Data":"d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3"} Mar 09 09:45:01 crc kubenswrapper[4692]: I0309 09:45:01.026067 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf"] Mar 09 09:45:01 crc kubenswrapper[4692]: I0309 09:45:01.736755 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpv7t" event={"ID":"96f563ba-b297-4ba0-9747-9a44c127c197","Type":"ContainerStarted","Data":"7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706"} Mar 09 09:45:01 crc kubenswrapper[4692]: I0309 09:45:01.739415 4692 generic.go:334] "Generic (PLEG): container finished" podID="c316cf6d-8a6f-46fd-a609-9bae80441c51" containerID="787e095e40a2c3cba35976f1e664893755d50fbf5a916d2b028aeeebc8c8e432" exitCode=0 Mar 09 09:45:01 crc kubenswrapper[4692]: I0309 09:45:01.739452 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" event={"ID":"c316cf6d-8a6f-46fd-a609-9bae80441c51","Type":"ContainerDied","Data":"787e095e40a2c3cba35976f1e664893755d50fbf5a916d2b028aeeebc8c8e432"} Mar 09 09:45:01 crc kubenswrapper[4692]: I0309 09:45:01.739488 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" event={"ID":"c316cf6d-8a6f-46fd-a609-9bae80441c51","Type":"ContainerStarted","Data":"f4d39cbda31e1bf02ac4209eff489456f5968a2d610991d48da5b2f76f3fcdf4"} Mar 09 09:45:01 crc kubenswrapper[4692]: I0309 09:45:01.778188 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lpv7t" podStartSLOduration=3.287718452 podStartE2EDuration="5.778143917s" podCreationTimestamp="2026-03-09 09:44:56 +0000 UTC" firstStartedPulling="2026-03-09 09:44:58.710861503 +0000 UTC m=+1499.535597094" lastFinishedPulling="2026-03-09 09:45:01.201286978 +0000 UTC m=+1502.026022559" observedRunningTime="2026-03-09 09:45:01.758138544 +0000 UTC m=+1502.582874135" watchObservedRunningTime="2026-03-09 09:45:01.778143917 +0000 UTC m=+1502.602879498" Mar 09 09:45:02 crc kubenswrapper[4692]: I0309 09:45:02.862512 4692 scope.go:117] "RemoveContainer" containerID="c20501be9e7f093efa0f3a1c0392cc53a297a64c0764c25306b40378d49d08e5" Mar 09 09:45:02 crc kubenswrapper[4692]: I0309 09:45:02.894661 4692 scope.go:117] "RemoveContainer" containerID="f7f92cd200436cefbfd944307e7af99d76ac4b846f0aad05045381c2845036e5" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.048621 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.207787 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c316cf6d-8a6f-46fd-a609-9bae80441c51-config-volume\") pod \"c316cf6d-8a6f-46fd-a609-9bae80441c51\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.207856 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6j2j\" (UniqueName: \"kubernetes.io/projected/c316cf6d-8a6f-46fd-a609-9bae80441c51-kube-api-access-m6j2j\") pod \"c316cf6d-8a6f-46fd-a609-9bae80441c51\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.207903 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c316cf6d-8a6f-46fd-a609-9bae80441c51-secret-volume\") pod \"c316cf6d-8a6f-46fd-a609-9bae80441c51\" (UID: \"c316cf6d-8a6f-46fd-a609-9bae80441c51\") " Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.208532 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c316cf6d-8a6f-46fd-a609-9bae80441c51-config-volume" (OuterVolumeSpecName: "config-volume") pod "c316cf6d-8a6f-46fd-a609-9bae80441c51" (UID: "c316cf6d-8a6f-46fd-a609-9bae80441c51"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.215491 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c316cf6d-8a6f-46fd-a609-9bae80441c51-kube-api-access-m6j2j" (OuterVolumeSpecName: "kube-api-access-m6j2j") pod "c316cf6d-8a6f-46fd-a609-9bae80441c51" (UID: "c316cf6d-8a6f-46fd-a609-9bae80441c51"). InnerVolumeSpecName "kube-api-access-m6j2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.215485 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c316cf6d-8a6f-46fd-a609-9bae80441c51-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c316cf6d-8a6f-46fd-a609-9bae80441c51" (UID: "c316cf6d-8a6f-46fd-a609-9bae80441c51"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.309546 4692 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c316cf6d-8a6f-46fd-a609-9bae80441c51-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.309574 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6j2j\" (UniqueName: \"kubernetes.io/projected/c316cf6d-8a6f-46fd-a609-9bae80441c51-kube-api-access-m6j2j\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.309585 4692 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c316cf6d-8a6f-46fd-a609-9bae80441c51-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.754788 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" event={"ID":"c316cf6d-8a6f-46fd-a609-9bae80441c51","Type":"ContainerDied","Data":"f4d39cbda31e1bf02ac4209eff489456f5968a2d610991d48da5b2f76f3fcdf4"} Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.754824 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4d39cbda31e1bf02ac4209eff489456f5968a2d610991d48da5b2f76f3fcdf4" Mar 09 09:45:03 crc kubenswrapper[4692]: I0309 09:45:03.754862 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550825-7rsqf" Mar 09 09:45:07 crc kubenswrapper[4692]: I0309 09:45:07.081644 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:45:07 crc kubenswrapper[4692]: I0309 09:45:07.083142 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:45:07 crc kubenswrapper[4692]: I0309 09:45:07.131842 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:45:07 crc kubenswrapper[4692]: I0309 09:45:07.824982 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:45:07 crc kubenswrapper[4692]: I0309 09:45:07.863506 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lpv7t"] Mar 09 09:45:09 crc kubenswrapper[4692]: I0309 09:45:09.796929 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lpv7t" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="registry-server" containerID="cri-o://7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706" gracePeriod=2 Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.209619 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.328612 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-catalog-content\") pod \"96f563ba-b297-4ba0-9747-9a44c127c197\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.328709 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-utilities\") pod \"96f563ba-b297-4ba0-9747-9a44c127c197\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.328756 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlbqf\" (UniqueName: \"kubernetes.io/projected/96f563ba-b297-4ba0-9747-9a44c127c197-kube-api-access-xlbqf\") pod \"96f563ba-b297-4ba0-9747-9a44c127c197\" (UID: \"96f563ba-b297-4ba0-9747-9a44c127c197\") " Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.329527 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-utilities" (OuterVolumeSpecName: "utilities") pod "96f563ba-b297-4ba0-9747-9a44c127c197" (UID: "96f563ba-b297-4ba0-9747-9a44c127c197"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.334420 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96f563ba-b297-4ba0-9747-9a44c127c197-kube-api-access-xlbqf" (OuterVolumeSpecName: "kube-api-access-xlbqf") pod "96f563ba-b297-4ba0-9747-9a44c127c197" (UID: "96f563ba-b297-4ba0-9747-9a44c127c197"). InnerVolumeSpecName "kube-api-access-xlbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.430466 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlbqf\" (UniqueName: \"kubernetes.io/projected/96f563ba-b297-4ba0-9747-9a44c127c197-kube-api-access-xlbqf\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.430506 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.617365 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96f563ba-b297-4ba0-9747-9a44c127c197" (UID: "96f563ba-b297-4ba0-9747-9a44c127c197"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.633341 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96f563ba-b297-4ba0-9747-9a44c127c197-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.806093 4692 generic.go:334] "Generic (PLEG): container finished" podID="96f563ba-b297-4ba0-9747-9a44c127c197" containerID="7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706" exitCode=0 Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.806152 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lpv7t" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.806153 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpv7t" event={"ID":"96f563ba-b297-4ba0-9747-9a44c127c197","Type":"ContainerDied","Data":"7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706"} Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.806237 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lpv7t" event={"ID":"96f563ba-b297-4ba0-9747-9a44c127c197","Type":"ContainerDied","Data":"6ed798fda3f1a4e475bb4e00352f3245211d17d5fc2504d41c21963da08feaf7"} Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.806263 4692 scope.go:117] "RemoveContainer" containerID="7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.832801 4692 scope.go:117] "RemoveContainer" containerID="d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.838411 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lpv7t"] Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.846058 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lpv7t"] Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.861036 4692 scope.go:117] "RemoveContainer" containerID="ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.890753 4692 scope.go:117] "RemoveContainer" containerID="7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706" Mar 09 09:45:10 crc kubenswrapper[4692]: E0309 09:45:10.891302 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706\": container with ID starting with 7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706 not found: ID does not exist" containerID="7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.891341 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706"} err="failed to get container status \"7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706\": rpc error: code = NotFound desc = could not find container \"7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706\": container with ID starting with 7d37b2b9f1567c9d7e1f11523dd2c495e9f85eeea8f3d8ac18247480ac9fa706 not found: ID does not exist" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.891372 4692 scope.go:117] "RemoveContainer" containerID="d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3" Mar 09 09:45:10 crc kubenswrapper[4692]: E0309 09:45:10.891599 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3\": container with ID starting with d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3 not found: ID does not exist" containerID="d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.891639 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3"} err="failed to get container status \"d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3\": rpc error: code = NotFound desc = could not find container \"d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3\": container with ID starting with d05fe59ac3cd3bcfb4d41aaa2f492ef0769090004ee39715dd68da989ca80aa3 not found: ID does not exist" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.891658 4692 scope.go:117] "RemoveContainer" containerID="ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81" Mar 09 09:45:10 crc kubenswrapper[4692]: E0309 09:45:10.891876 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81\": container with ID starting with ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81 not found: ID does not exist" containerID="ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81" Mar 09 09:45:10 crc kubenswrapper[4692]: I0309 09:45:10.891903 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81"} err="failed to get container status \"ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81\": rpc error: code = NotFound desc = could not find container \"ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81\": container with ID starting with ee6cf114c62268372b5952f1f80bd472d4f9a28118b182f5b262cb5b91d2fe81 not found: ID does not exist" Mar 09 09:45:12 crc kubenswrapper[4692]: I0309 09:45:12.079818 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" path="/var/lib/kubelet/pods/96f563ba-b297-4ba0-9747-9a44c127c197/volumes" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.624746 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.625311 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.848241 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nkfrd"] Mar 09 09:45:17 crc kubenswrapper[4692]: E0309 09:45:17.848866 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="registry-server" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.848887 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="registry-server" Mar 09 09:45:17 crc kubenswrapper[4692]: E0309 09:45:17.848906 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="extract-content" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.848914 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="extract-content" Mar 09 09:45:17 crc kubenswrapper[4692]: E0309 09:45:17.848944 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="extract-utilities" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.848953 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="extract-utilities" Mar 09 09:45:17 crc kubenswrapper[4692]: E0309 09:45:17.848989 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c316cf6d-8a6f-46fd-a609-9bae80441c51" containerName="collect-profiles" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.848997 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c316cf6d-8a6f-46fd-a609-9bae80441c51" containerName="collect-profiles" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.849376 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="96f563ba-b297-4ba0-9747-9a44c127c197" containerName="registry-server" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.849424 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c316cf6d-8a6f-46fd-a609-9bae80441c51" containerName="collect-profiles" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.853952 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.877555 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkfrd"] Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.945613 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8gqk\" (UniqueName: \"kubernetes.io/projected/d4fca79b-6a1e-4af4-9445-be3469c3b65b-kube-api-access-m8gqk\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.946086 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-catalog-content\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:17 crc kubenswrapper[4692]: I0309 09:45:17.946345 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-utilities\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.047540 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-utilities\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.047630 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8gqk\" (UniqueName: \"kubernetes.io/projected/d4fca79b-6a1e-4af4-9445-be3469c3b65b-kube-api-access-m8gqk\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.047677 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-catalog-content\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.048132 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-utilities\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.048148 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-catalog-content\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.066819 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8gqk\" (UniqueName: \"kubernetes.io/projected/d4fca79b-6a1e-4af4-9445-be3469c3b65b-kube-api-access-m8gqk\") pod \"redhat-marketplace-nkfrd\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.194291 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.616323 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkfrd"] Mar 09 09:45:18 crc kubenswrapper[4692]: I0309 09:45:18.881153 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkfrd" event={"ID":"d4fca79b-6a1e-4af4-9445-be3469c3b65b","Type":"ContainerStarted","Data":"c87b2eb7020931d9c4e661cc513b73640946d82478bdc2ca742fa2382974eef5"} Mar 09 09:45:19 crc kubenswrapper[4692]: I0309 09:45:19.889477 4692 generic.go:334] "Generic (PLEG): container finished" podID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerID="4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b" exitCode=0 Mar 09 09:45:19 crc kubenswrapper[4692]: I0309 09:45:19.889520 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkfrd" event={"ID":"d4fca79b-6a1e-4af4-9445-be3469c3b65b","Type":"ContainerDied","Data":"4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b"} Mar 09 09:45:20 crc kubenswrapper[4692]: I0309 09:45:20.898590 4692 generic.go:334] "Generic (PLEG): container finished" podID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerID="05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6" exitCode=0 Mar 09 09:45:20 crc kubenswrapper[4692]: I0309 09:45:20.898703 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkfrd" event={"ID":"d4fca79b-6a1e-4af4-9445-be3469c3b65b","Type":"ContainerDied","Data":"05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6"} Mar 09 09:45:21 crc kubenswrapper[4692]: I0309 09:45:21.907678 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkfrd" event={"ID":"d4fca79b-6a1e-4af4-9445-be3469c3b65b","Type":"ContainerStarted","Data":"bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c"} Mar 09 09:45:21 crc kubenswrapper[4692]: I0309 09:45:21.927309 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nkfrd" podStartSLOduration=3.561572696 podStartE2EDuration="4.927292969s" podCreationTimestamp="2026-03-09 09:45:17 +0000 UTC" firstStartedPulling="2026-03-09 09:45:19.891663879 +0000 UTC m=+1520.716399460" lastFinishedPulling="2026-03-09 09:45:21.257384152 +0000 UTC m=+1522.082119733" observedRunningTime="2026-03-09 09:45:21.921948748 +0000 UTC m=+1522.746684349" watchObservedRunningTime="2026-03-09 09:45:21.927292969 +0000 UTC m=+1522.752028550" Mar 09 09:45:28 crc kubenswrapper[4692]: I0309 09:45:28.195336 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:28 crc kubenswrapper[4692]: I0309 09:45:28.196038 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:28 crc kubenswrapper[4692]: I0309 09:45:28.245361 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:29 crc kubenswrapper[4692]: I0309 09:45:29.003613 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:29 crc kubenswrapper[4692]: I0309 09:45:29.044812 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkfrd"] Mar 09 09:45:30 crc kubenswrapper[4692]: I0309 09:45:30.970130 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nkfrd" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="registry-server" containerID="cri-o://bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c" gracePeriod=2 Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.432844 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.545813 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-utilities\") pod \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.545901 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-catalog-content\") pod \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.546033 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8gqk\" (UniqueName: \"kubernetes.io/projected/d4fca79b-6a1e-4af4-9445-be3469c3b65b-kube-api-access-m8gqk\") pod \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\" (UID: \"d4fca79b-6a1e-4af4-9445-be3469c3b65b\") " Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.547149 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-utilities" (OuterVolumeSpecName: "utilities") pod "d4fca79b-6a1e-4af4-9445-be3469c3b65b" (UID: "d4fca79b-6a1e-4af4-9445-be3469c3b65b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.564415 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4fca79b-6a1e-4af4-9445-be3469c3b65b-kube-api-access-m8gqk" (OuterVolumeSpecName: "kube-api-access-m8gqk") pod "d4fca79b-6a1e-4af4-9445-be3469c3b65b" (UID: "d4fca79b-6a1e-4af4-9445-be3469c3b65b"). InnerVolumeSpecName "kube-api-access-m8gqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.574607 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4fca79b-6a1e-4af4-9445-be3469c3b65b" (UID: "d4fca79b-6a1e-4af4-9445-be3469c3b65b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.647939 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.647976 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8gqk\" (UniqueName: \"kubernetes.io/projected/d4fca79b-6a1e-4af4-9445-be3469c3b65b-kube-api-access-m8gqk\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.647989 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4fca79b-6a1e-4af4-9445-be3469c3b65b-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.981515 4692 generic.go:334] "Generic (PLEG): container finished" podID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerID="bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c" exitCode=0 Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.981568 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkfrd" event={"ID":"d4fca79b-6a1e-4af4-9445-be3469c3b65b","Type":"ContainerDied","Data":"bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c"} Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.981580 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nkfrd" Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.981598 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nkfrd" event={"ID":"d4fca79b-6a1e-4af4-9445-be3469c3b65b","Type":"ContainerDied","Data":"c87b2eb7020931d9c4e661cc513b73640946d82478bdc2ca742fa2382974eef5"} Mar 09 09:45:31 crc kubenswrapper[4692]: I0309 09:45:31.981619 4692 scope.go:117] "RemoveContainer" containerID="bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.007629 4692 scope.go:117] "RemoveContainer" containerID="05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.032849 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkfrd"] Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.033896 4692 scope.go:117] "RemoveContainer" containerID="4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.038314 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nkfrd"] Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.061016 4692 scope.go:117] "RemoveContainer" containerID="bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c" Mar 09 09:45:32 crc kubenswrapper[4692]: E0309 09:45:32.061690 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c\": container with ID starting with bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c not found: ID does not exist" containerID="bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.061756 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c"} err="failed to get container status \"bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c\": rpc error: code = NotFound desc = could not find container \"bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c\": container with ID starting with bf20be543d66ba3525f5c3adc8b88438ad75900816ef88ef3a4cbf87a603577c not found: ID does not exist" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.061782 4692 scope.go:117] "RemoveContainer" containerID="05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6" Mar 09 09:45:32 crc kubenswrapper[4692]: E0309 09:45:32.062146 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6\": container with ID starting with 05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6 not found: ID does not exist" containerID="05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.062201 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6"} err="failed to get container status \"05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6\": rpc error: code = NotFound desc = could not find container \"05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6\": container with ID starting with 05b8febcf018c7c76a2cdd88c665ef4eb056a3cd8d6b2030e31995614a2cb2f6 not found: ID does not exist" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.062215 4692 scope.go:117] "RemoveContainer" containerID="4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b" Mar 09 09:45:32 crc kubenswrapper[4692]: E0309 09:45:32.062492 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b\": container with ID starting with 4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b not found: ID does not exist" containerID="4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.062515 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b"} err="failed to get container status \"4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b\": rpc error: code = NotFound desc = could not find container \"4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b\": container with ID starting with 4eec3d837152d8efd8f9e4b061433141128a33bc198ab5b29e5f02b2f690570b not found: ID does not exist" Mar 09 09:45:32 crc kubenswrapper[4692]: I0309 09:45:32.080481 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" path="/var/lib/kubelet/pods/d4fca79b-6a1e-4af4-9445-be3469c3b65b/volumes" Mar 09 09:45:47 crc kubenswrapper[4692]: I0309 09:45:47.623687 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:45:47 crc kubenswrapper[4692]: I0309 09:45:47.624268 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:45:47 crc kubenswrapper[4692]: I0309 09:45:47.624320 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:45:47 crc kubenswrapper[4692]: I0309 09:45:47.624984 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:45:47 crc kubenswrapper[4692]: I0309 09:45:47.625054 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" gracePeriod=600 Mar 09 09:45:47 crc kubenswrapper[4692]: E0309 09:45:47.756886 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:45:48 crc kubenswrapper[4692]: I0309 09:45:48.085720 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" exitCode=0 Mar 09 09:45:48 crc kubenswrapper[4692]: I0309 09:45:48.085766 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46"} Mar 09 09:45:48 crc kubenswrapper[4692]: I0309 09:45:48.085842 4692 scope.go:117] "RemoveContainer" containerID="dc5528f2b79383c92cd6d5389f75e574b39b690b0caddba49d4c6810fb3c85bb" Mar 09 09:45:48 crc kubenswrapper[4692]: I0309 09:45:48.086351 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:45:48 crc kubenswrapper[4692]: E0309 09:45:48.086675 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:45:53 crc kubenswrapper[4692]: I0309 09:45:53.055536 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/root-account-create-update-xp2sq"] Mar 09 09:45:53 crc kubenswrapper[4692]: I0309 09:45:53.061017 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/root-account-create-update-xp2sq"] Mar 09 09:45:54 crc kubenswrapper[4692]: I0309 09:45:54.079986 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1" path="/var/lib/kubelet/pods/d5cc4b5f-719e-4b10-8ea5-53f49e0db3b1/volumes" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.138877 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550826-9xb8q"] Mar 09 09:46:00 crc kubenswrapper[4692]: E0309 09:46:00.139824 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="registry-server" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.139840 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="registry-server" Mar 09 09:46:00 crc kubenswrapper[4692]: E0309 09:46:00.139861 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="extract-content" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.139868 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="extract-content" Mar 09 09:46:00 crc kubenswrapper[4692]: E0309 09:46:00.139883 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="extract-utilities" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.139890 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="extract-utilities" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.140112 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4fca79b-6a1e-4af4-9445-be3469c3b65b" containerName="registry-server" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.140764 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.143039 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.143114 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.143282 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.148741 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz6nk\" (UniqueName: \"kubernetes.io/projected/d4684029-a5c7-40e6-9d76-db09d57818e8-kube-api-access-qz6nk\") pod \"auto-csr-approver-29550826-9xb8q\" (UID: \"d4684029-a5c7-40e6-9d76-db09d57818e8\") " pod="openshift-infra/auto-csr-approver-29550826-9xb8q" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.154308 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550826-9xb8q"] Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.250651 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz6nk\" (UniqueName: \"kubernetes.io/projected/d4684029-a5c7-40e6-9d76-db09d57818e8-kube-api-access-qz6nk\") pod \"auto-csr-approver-29550826-9xb8q\" (UID: \"d4684029-a5c7-40e6-9d76-db09d57818e8\") " pod="openshift-infra/auto-csr-approver-29550826-9xb8q" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.269688 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz6nk\" (UniqueName: \"kubernetes.io/projected/d4684029-a5c7-40e6-9d76-db09d57818e8-kube-api-access-qz6nk\") pod \"auto-csr-approver-29550826-9xb8q\" (UID: \"d4684029-a5c7-40e6-9d76-db09d57818e8\") " pod="openshift-infra/auto-csr-approver-29550826-9xb8q" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.461149 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" Mar 09 09:46:00 crc kubenswrapper[4692]: I0309 09:46:00.867528 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550826-9xb8q"] Mar 09 09:46:01 crc kubenswrapper[4692]: I0309 09:46:01.174635 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" event={"ID":"d4684029-a5c7-40e6-9d76-db09d57818e8","Type":"ContainerStarted","Data":"6042a2d8a5ce08c8d5b334f102ab3ee39706e8aa2912e3d263d0e5b7c1582c92"} Mar 09 09:46:02 crc kubenswrapper[4692]: I0309 09:46:02.071902 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:46:02 crc kubenswrapper[4692]: E0309 09:46:02.072453 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:46:02 crc kubenswrapper[4692]: I0309 09:46:02.184228 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" event={"ID":"d4684029-a5c7-40e6-9d76-db09d57818e8","Type":"ContainerStarted","Data":"40d197987b30e35706f6500edf302b8d02e84068e88e2caabb4b823448007189"} Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.006303 4692 scope.go:117] "RemoveContainer" containerID="704ea831492a29f5a461f465943223a5cf5a1275ba375cc94f8c0af1c8c9c7dc" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.033692 4692 scope.go:117] "RemoveContainer" containerID="d9fca122f4493bee53c9b8f344da247b633bac2de061a56c8724e98fc1516460" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.054548 4692 scope.go:117] "RemoveContainer" containerID="8eabfccd02936e226e1669133e8352835a91c3b973895cb307ea66b9d7fa12d1" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.076915 4692 scope.go:117] "RemoveContainer" containerID="4dcd159606ea3eb0a56b35c2bbc8955b8cf4ee1aeaddc4a57f92fb1b5c579662" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.103827 4692 scope.go:117] "RemoveContainer" containerID="ccb5034272c4443a5d30c3e77c9d8499de513d71548f5dfc4aa471c4cf783678" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.124316 4692 scope.go:117] "RemoveContainer" containerID="cbf6beff918423c3958f88e8749e875ff0a036cc342ad42299854971e4745a5a" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.142048 4692 scope.go:117] "RemoveContainer" containerID="0574622c3ea59457671ae2d3e5912c7907129a0385b9617eb6c635283f6aedf8" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.157256 4692 scope.go:117] "RemoveContainer" containerID="7d475f88fe15d82cbca79a9f0f44b9547c75bd6a08ec54a080ad68c5c9eb1667" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.173316 4692 scope.go:117] "RemoveContainer" containerID="26c226cab26d728919ade2a94001899645db4128cc63122ef0b7e131dafe18e4" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.191352 4692 scope.go:117] "RemoveContainer" containerID="138c618fb48aac30a1bb3124b9eb5bd1f25f2df24b7e57329768d7535fad9638" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.196792 4692 generic.go:334] "Generic (PLEG): container finished" podID="d4684029-a5c7-40e6-9d76-db09d57818e8" containerID="40d197987b30e35706f6500edf302b8d02e84068e88e2caabb4b823448007189" exitCode=0 Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.196864 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" event={"ID":"d4684029-a5c7-40e6-9d76-db09d57818e8","Type":"ContainerDied","Data":"40d197987b30e35706f6500edf302b8d02e84068e88e2caabb4b823448007189"} Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.215456 4692 scope.go:117] "RemoveContainer" containerID="dfc762ad83fc57e117251ed04fa58e82ee932457ae195cc07a45d8582c6f48b3" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.483326 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.498487 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz6nk\" (UniqueName: \"kubernetes.io/projected/d4684029-a5c7-40e6-9d76-db09d57818e8-kube-api-access-qz6nk\") pod \"d4684029-a5c7-40e6-9d76-db09d57818e8\" (UID: \"d4684029-a5c7-40e6-9d76-db09d57818e8\") " Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.517662 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4684029-a5c7-40e6-9d76-db09d57818e8-kube-api-access-qz6nk" (OuterVolumeSpecName: "kube-api-access-qz6nk") pod "d4684029-a5c7-40e6-9d76-db09d57818e8" (UID: "d4684029-a5c7-40e6-9d76-db09d57818e8"). InnerVolumeSpecName "kube-api-access-qz6nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:46:03 crc kubenswrapper[4692]: I0309 09:46:03.600026 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz6nk\" (UniqueName: \"kubernetes.io/projected/d4684029-a5c7-40e6-9d76-db09d57818e8-kube-api-access-qz6nk\") on node \"crc\" DevicePath \"\"" Mar 09 09:46:04 crc kubenswrapper[4692]: I0309 09:46:04.209502 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" event={"ID":"d4684029-a5c7-40e6-9d76-db09d57818e8","Type":"ContainerDied","Data":"6042a2d8a5ce08c8d5b334f102ab3ee39706e8aa2912e3d263d0e5b7c1582c92"} Mar 09 09:46:04 crc kubenswrapper[4692]: I0309 09:46:04.209559 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6042a2d8a5ce08c8d5b334f102ab3ee39706e8aa2912e3d263d0e5b7c1582c92" Mar 09 09:46:04 crc kubenswrapper[4692]: I0309 09:46:04.209569 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550826-9xb8q" Mar 09 09:46:04 crc kubenswrapper[4692]: I0309 09:46:04.550136 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550820-jnhqp"] Mar 09 09:46:04 crc kubenswrapper[4692]: I0309 09:46:04.555158 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550820-jnhqp"] Mar 09 09:46:06 crc kubenswrapper[4692]: I0309 09:46:06.079702 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4193ac81-13f3-42db-afd3-bbc890b0000f" path="/var/lib/kubelet/pods/4193ac81-13f3-42db-afd3-bbc890b0000f/volumes" Mar 09 09:46:13 crc kubenswrapper[4692]: I0309 09:46:13.072687 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:46:13 crc kubenswrapper[4692]: E0309 09:46:13.073388 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:46:28 crc kubenswrapper[4692]: I0309 09:46:28.071322 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:46:28 crc kubenswrapper[4692]: E0309 09:46:28.071915 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:46:39 crc kubenswrapper[4692]: I0309 09:46:39.071947 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:46:39 crc kubenswrapper[4692]: E0309 09:46:39.072603 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:46:50 crc kubenswrapper[4692]: I0309 09:46:50.077360 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:46:50 crc kubenswrapper[4692]: E0309 09:46:50.078133 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:47:01 crc kubenswrapper[4692]: I0309 09:47:01.071816 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:47:01 crc kubenswrapper[4692]: E0309 09:47:01.072518 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:47:03 crc kubenswrapper[4692]: I0309 09:47:03.311838 4692 scope.go:117] "RemoveContainer" containerID="5ee28c43242d8b55f00a22bcffc9a222f8c0cec8d60d2e6f93410be897772cfb" Mar 09 09:47:03 crc kubenswrapper[4692]: I0309 09:47:03.349102 4692 scope.go:117] "RemoveContainer" containerID="74e1a884f7f9f8904a07b1b8011e4ba8e3c1816f8dcff5525e38a3d54150a0b7" Mar 09 09:47:15 crc kubenswrapper[4692]: I0309 09:47:15.071802 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:47:15 crc kubenswrapper[4692]: E0309 09:47:15.072547 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:47:29 crc kubenswrapper[4692]: I0309 09:47:29.035803 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-db-create-md87c"] Mar 09 09:47:29 crc kubenswrapper[4692]: I0309 09:47:29.042380 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-db-create-md87c"] Mar 09 09:47:29 crc kubenswrapper[4692]: I0309 09:47:29.065066 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-cb07-account-create-update-sknx7"] Mar 09 09:47:29 crc kubenswrapper[4692]: I0309 09:47:29.071761 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:47:29 crc kubenswrapper[4692]: E0309 09:47:29.072045 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:47:29 crc kubenswrapper[4692]: I0309 09:47:29.073498 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-cb07-account-create-update-sknx7"] Mar 09 09:47:30 crc kubenswrapper[4692]: I0309 09:47:30.082934 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c67531d0-eb39-4d97-aea1-4ff7c33fc191" path="/var/lib/kubelet/pods/c67531d0-eb39-4d97-aea1-4ff7c33fc191/volumes" Mar 09 09:47:30 crc kubenswrapper[4692]: I0309 09:47:30.083773 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc" path="/var/lib/kubelet/pods/e8ed2661-3d2a-4611-bb6f-ee15d5b0f6cc/volumes" Mar 09 09:47:40 crc kubenswrapper[4692]: I0309 09:47:40.077020 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:47:40 crc kubenswrapper[4692]: E0309 09:47:40.078871 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:47:45 crc kubenswrapper[4692]: I0309 09:47:45.039386 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-l8bgh"] Mar 09 09:47:45 crc kubenswrapper[4692]: I0309 09:47:45.046011 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-db-sync-l8bgh"] Mar 09 09:47:46 crc kubenswrapper[4692]: I0309 09:47:46.079551 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba1c1067-d220-48ff-b60c-d0947ee5b1d9" path="/var/lib/kubelet/pods/ba1c1067-d220-48ff-b60c-d0947ee5b1d9/volumes" Mar 09 09:47:51 crc kubenswrapper[4692]: I0309 09:47:51.023597 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-pg4dm"] Mar 09 09:47:51 crc kubenswrapper[4692]: I0309 09:47:51.030256 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/keystone-bootstrap-pg4dm"] Mar 09 09:47:52 crc kubenswrapper[4692]: I0309 09:47:52.072499 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:47:52 crc kubenswrapper[4692]: E0309 09:47:52.072962 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:47:52 crc kubenswrapper[4692]: I0309 09:47:52.085120 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f420a643-bfec-4568-8785-b1b18c1cf320" path="/var/lib/kubelet/pods/f420a643-bfec-4568-8785-b1b18c1cf320/volumes" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.147598 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550828-dthsk"] Mar 09 09:48:00 crc kubenswrapper[4692]: E0309 09:48:00.148401 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4684029-a5c7-40e6-9d76-db09d57818e8" containerName="oc" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.148414 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4684029-a5c7-40e6-9d76-db09d57818e8" containerName="oc" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.148534 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4684029-a5c7-40e6-9d76-db09d57818e8" containerName="oc" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.148956 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550828-dthsk" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.155451 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.155669 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.157585 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.162259 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550828-dthsk"] Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.299731 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp8v5\" (UniqueName: \"kubernetes.io/projected/3dd6688d-ce93-4f18-8b81-17f28509fe24-kube-api-access-hp8v5\") pod \"auto-csr-approver-29550828-dthsk\" (UID: \"3dd6688d-ce93-4f18-8b81-17f28509fe24\") " pod="openshift-infra/auto-csr-approver-29550828-dthsk" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.400969 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp8v5\" (UniqueName: \"kubernetes.io/projected/3dd6688d-ce93-4f18-8b81-17f28509fe24-kube-api-access-hp8v5\") pod \"auto-csr-approver-29550828-dthsk\" (UID: \"3dd6688d-ce93-4f18-8b81-17f28509fe24\") " pod="openshift-infra/auto-csr-approver-29550828-dthsk" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.421465 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp8v5\" (UniqueName: \"kubernetes.io/projected/3dd6688d-ce93-4f18-8b81-17f28509fe24-kube-api-access-hp8v5\") pod \"auto-csr-approver-29550828-dthsk\" (UID: \"3dd6688d-ce93-4f18-8b81-17f28509fe24\") " pod="openshift-infra/auto-csr-approver-29550828-dthsk" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.467909 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550828-dthsk" Mar 09 09:48:00 crc kubenswrapper[4692]: I0309 09:48:00.891393 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550828-dthsk"] Mar 09 09:48:00 crc kubenswrapper[4692]: W0309 09:48:00.899573 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dd6688d_ce93_4f18_8b81_17f28509fe24.slice/crio-37b059e1eebb8eedeb8629d1e1f6f7820ce3a654806e3f6e470f6bbe7c18e74f WatchSource:0}: Error finding container 37b059e1eebb8eedeb8629d1e1f6f7820ce3a654806e3f6e470f6bbe7c18e74f: Status 404 returned error can't find the container with id 37b059e1eebb8eedeb8629d1e1f6f7820ce3a654806e3f6e470f6bbe7c18e74f Mar 09 09:48:01 crc kubenswrapper[4692]: I0309 09:48:01.018059 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550828-dthsk" event={"ID":"3dd6688d-ce93-4f18-8b81-17f28509fe24","Type":"ContainerStarted","Data":"37b059e1eebb8eedeb8629d1e1f6f7820ce3a654806e3f6e470f6bbe7c18e74f"} Mar 09 09:48:03 crc kubenswrapper[4692]: I0309 09:48:03.038211 4692 generic.go:334] "Generic (PLEG): container finished" podID="3dd6688d-ce93-4f18-8b81-17f28509fe24" containerID="54d9bba991c4580b8ebc695ea147000337f9dc690409245583b6f21349e8c8b7" exitCode=0 Mar 09 09:48:03 crc kubenswrapper[4692]: I0309 09:48:03.038261 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550828-dthsk" event={"ID":"3dd6688d-ce93-4f18-8b81-17f28509fe24","Type":"ContainerDied","Data":"54d9bba991c4580b8ebc695ea147000337f9dc690409245583b6f21349e8c8b7"} Mar 09 09:48:03 crc kubenswrapper[4692]: I0309 09:48:03.415906 4692 scope.go:117] "RemoveContainer" containerID="87872a7b618c5191831e36a7519ff32436d180558935d34db116b477921f497c" Mar 09 09:48:03 crc kubenswrapper[4692]: I0309 09:48:03.491059 4692 scope.go:117] "RemoveContainer" containerID="138283a861b96ca4ff869cad6bc6e67d44a003b52bddd47c62434dce3f45ebd5" Mar 09 09:48:03 crc kubenswrapper[4692]: I0309 09:48:03.513295 4692 scope.go:117] "RemoveContainer" containerID="7ea425a2b3f008a6d3f1042de8b2947a69718e271a3a4348a6c56f9e8fa2f5b8" Mar 09 09:48:03 crc kubenswrapper[4692]: I0309 09:48:03.547283 4692 scope.go:117] "RemoveContainer" containerID="cc28cb1a5e009ae73ba2700723be66b48e4d67e1c1d76931d82c683f007bfd38" Mar 09 09:48:03 crc kubenswrapper[4692]: I0309 09:48:03.589636 4692 scope.go:117] "RemoveContainer" containerID="536fe1e790da3e170a5692de5f9cf6ab816b621b035dca5db8bd6ecf2190efdc" Mar 09 09:48:04 crc kubenswrapper[4692]: I0309 09:48:04.251374 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550828-dthsk" Mar 09 09:48:04 crc kubenswrapper[4692]: I0309 09:48:04.352402 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp8v5\" (UniqueName: \"kubernetes.io/projected/3dd6688d-ce93-4f18-8b81-17f28509fe24-kube-api-access-hp8v5\") pod \"3dd6688d-ce93-4f18-8b81-17f28509fe24\" (UID: \"3dd6688d-ce93-4f18-8b81-17f28509fe24\") " Mar 09 09:48:04 crc kubenswrapper[4692]: I0309 09:48:04.358681 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd6688d-ce93-4f18-8b81-17f28509fe24-kube-api-access-hp8v5" (OuterVolumeSpecName: "kube-api-access-hp8v5") pod "3dd6688d-ce93-4f18-8b81-17f28509fe24" (UID: "3dd6688d-ce93-4f18-8b81-17f28509fe24"). InnerVolumeSpecName "kube-api-access-hp8v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:48:04 crc kubenswrapper[4692]: I0309 09:48:04.454482 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp8v5\" (UniqueName: \"kubernetes.io/projected/3dd6688d-ce93-4f18-8b81-17f28509fe24-kube-api-access-hp8v5\") on node \"crc\" DevicePath \"\"" Mar 09 09:48:05 crc kubenswrapper[4692]: I0309 09:48:05.051914 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550828-dthsk" event={"ID":"3dd6688d-ce93-4f18-8b81-17f28509fe24","Type":"ContainerDied","Data":"37b059e1eebb8eedeb8629d1e1f6f7820ce3a654806e3f6e470f6bbe7c18e74f"} Mar 09 09:48:05 crc kubenswrapper[4692]: I0309 09:48:05.051952 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37b059e1eebb8eedeb8629d1e1f6f7820ce3a654806e3f6e470f6bbe7c18e74f" Mar 09 09:48:05 crc kubenswrapper[4692]: I0309 09:48:05.051961 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550828-dthsk" Mar 09 09:48:05 crc kubenswrapper[4692]: I0309 09:48:05.071747 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:48:05 crc kubenswrapper[4692]: E0309 09:48:05.071959 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:48:05 crc kubenswrapper[4692]: I0309 09:48:05.303871 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550822-g4wkt"] Mar 09 09:48:05 crc kubenswrapper[4692]: I0309 09:48:05.308727 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550822-g4wkt"] Mar 09 09:48:06 crc kubenswrapper[4692]: I0309 09:48:06.080632 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="059e481e-f65b-4413-969e-edf4666c4e63" path="/var/lib/kubelet/pods/059e481e-f65b-4413-969e-edf4666c4e63/volumes" Mar 09 09:48:16 crc kubenswrapper[4692]: I0309 09:48:16.071866 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:48:16 crc kubenswrapper[4692]: E0309 09:48:16.072472 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:48:27 crc kubenswrapper[4692]: I0309 09:48:27.072473 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:48:27 crc kubenswrapper[4692]: E0309 09:48:27.073139 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:48:29 crc kubenswrapper[4692]: I0309 09:48:29.033713 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/barbican-db-create-z4kvk"] Mar 09 09:48:29 crc kubenswrapper[4692]: I0309 09:48:29.039721 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/barbican-990a-account-create-update-kcj5x"] Mar 09 09:48:29 crc kubenswrapper[4692]: I0309 09:48:29.044666 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/barbican-db-create-z4kvk"] Mar 09 09:48:29 crc kubenswrapper[4692]: I0309 09:48:29.049247 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/barbican-990a-account-create-update-kcj5x"] Mar 09 09:48:30 crc kubenswrapper[4692]: I0309 09:48:30.079447 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ff25a5a-862c-4367-a86d-3f6a5afec5a7" path="/var/lib/kubelet/pods/6ff25a5a-862c-4367-a86d-3f6a5afec5a7/volumes" Mar 09 09:48:30 crc kubenswrapper[4692]: I0309 09:48:30.080233 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d85ecf61-be76-4439-a305-d4a64fae30f3" path="/var/lib/kubelet/pods/d85ecf61-be76-4439-a305-d4a64fae30f3/volumes" Mar 09 09:48:33 crc kubenswrapper[4692]: E0309 09:48:33.182860 4692 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.181:43978->38.102.83.181:35901: write tcp 38.102.83.181:43978->38.102.83.181:35901: write: broken pipe Mar 09 09:48:38 crc kubenswrapper[4692]: I0309 09:48:38.071313 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:48:38 crc kubenswrapper[4692]: E0309 09:48:38.072047 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:48:52 crc kubenswrapper[4692]: I0309 09:48:52.072231 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:48:52 crc kubenswrapper[4692]: E0309 09:48:52.073016 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:49:03 crc kubenswrapper[4692]: I0309 09:49:03.663645 4692 scope.go:117] "RemoveContainer" containerID="8468b2f345979089bbe533ee7e8ea32f7e7d3806c32bb748fae2a1772cdd78b8" Mar 09 09:49:03 crc kubenswrapper[4692]: I0309 09:49:03.686464 4692 scope.go:117] "RemoveContainer" containerID="b6ee2239f970169d9a9900d51f829435dd2ed84e3c45b62a56a77b685c8e6700" Mar 09 09:49:03 crc kubenswrapper[4692]: I0309 09:49:03.709256 4692 scope.go:117] "RemoveContainer" containerID="200d9b7a50a0ab84887facd9cd4431b8973fc90227411f1b5e8252dbc9ebdc2c" Mar 09 09:49:07 crc kubenswrapper[4692]: I0309 09:49:07.071261 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:49:07 crc kubenswrapper[4692]: E0309 09:49:07.071581 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:49:22 crc kubenswrapper[4692]: I0309 09:49:22.081104 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:49:22 crc kubenswrapper[4692]: E0309 09:49:22.081969 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:49:33 crc kubenswrapper[4692]: I0309 09:49:33.071557 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:49:33 crc kubenswrapper[4692]: E0309 09:49:33.072624 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:49:47 crc kubenswrapper[4692]: I0309 09:49:47.071409 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:49:47 crc kubenswrapper[4692]: E0309 09:49:47.072132 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.135455 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550830-n66j7"] Mar 09 09:50:00 crc kubenswrapper[4692]: E0309 09:50:00.137918 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd6688d-ce93-4f18-8b81-17f28509fe24" containerName="oc" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.138026 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd6688d-ce93-4f18-8b81-17f28509fe24" containerName="oc" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.138344 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd6688d-ce93-4f18-8b81-17f28509fe24" containerName="oc" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.138952 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550830-n66j7" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.141495 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.142810 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.143836 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.150601 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550830-n66j7"] Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.300688 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l797p\" (UniqueName: \"kubernetes.io/projected/4bac7da7-e0f6-4875-b683-9dd6842c915b-kube-api-access-l797p\") pod \"auto-csr-approver-29550830-n66j7\" (UID: \"4bac7da7-e0f6-4875-b683-9dd6842c915b\") " pod="openshift-infra/auto-csr-approver-29550830-n66j7" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.402403 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l797p\" (UniqueName: \"kubernetes.io/projected/4bac7da7-e0f6-4875-b683-9dd6842c915b-kube-api-access-l797p\") pod \"auto-csr-approver-29550830-n66j7\" (UID: \"4bac7da7-e0f6-4875-b683-9dd6842c915b\") " pod="openshift-infra/auto-csr-approver-29550830-n66j7" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.422200 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l797p\" (UniqueName: \"kubernetes.io/projected/4bac7da7-e0f6-4875-b683-9dd6842c915b-kube-api-access-l797p\") pod \"auto-csr-approver-29550830-n66j7\" (UID: \"4bac7da7-e0f6-4875-b683-9dd6842c915b\") " pod="openshift-infra/auto-csr-approver-29550830-n66j7" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.457075 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550830-n66j7" Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.857465 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550830-n66j7"] Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.864856 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 09:50:00 crc kubenswrapper[4692]: I0309 09:50:00.889842 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550830-n66j7" event={"ID":"4bac7da7-e0f6-4875-b683-9dd6842c915b","Type":"ContainerStarted","Data":"d81de4dfe6952846acab2f7b1df19fb20feb5a135536a7478b7b6bfe7edc5207"} Mar 09 09:50:01 crc kubenswrapper[4692]: I0309 09:50:01.071711 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:50:01 crc kubenswrapper[4692]: E0309 09:50:01.071955 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:50:02 crc kubenswrapper[4692]: I0309 09:50:02.911737 4692 generic.go:334] "Generic (PLEG): container finished" podID="4bac7da7-e0f6-4875-b683-9dd6842c915b" containerID="005da3fb4a7d971c028d2a59b2a17104c914537c845235e63f1413e6f7bc3b8d" exitCode=0 Mar 09 09:50:02 crc kubenswrapper[4692]: I0309 09:50:02.911855 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550830-n66j7" event={"ID":"4bac7da7-e0f6-4875-b683-9dd6842c915b","Type":"ContainerDied","Data":"005da3fb4a7d971c028d2a59b2a17104c914537c845235e63f1413e6f7bc3b8d"} Mar 09 09:50:04 crc kubenswrapper[4692]: I0309 09:50:04.180332 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550830-n66j7" Mar 09 09:50:04 crc kubenswrapper[4692]: I0309 09:50:04.259326 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l797p\" (UniqueName: \"kubernetes.io/projected/4bac7da7-e0f6-4875-b683-9dd6842c915b-kube-api-access-l797p\") pod \"4bac7da7-e0f6-4875-b683-9dd6842c915b\" (UID: \"4bac7da7-e0f6-4875-b683-9dd6842c915b\") " Mar 09 09:50:04 crc kubenswrapper[4692]: I0309 09:50:04.265662 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bac7da7-e0f6-4875-b683-9dd6842c915b-kube-api-access-l797p" (OuterVolumeSpecName: "kube-api-access-l797p") pod "4bac7da7-e0f6-4875-b683-9dd6842c915b" (UID: "4bac7da7-e0f6-4875-b683-9dd6842c915b"). InnerVolumeSpecName "kube-api-access-l797p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:50:04 crc kubenswrapper[4692]: I0309 09:50:04.360536 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l797p\" (UniqueName: \"kubernetes.io/projected/4bac7da7-e0f6-4875-b683-9dd6842c915b-kube-api-access-l797p\") on node \"crc\" DevicePath \"\"" Mar 09 09:50:04 crc kubenswrapper[4692]: I0309 09:50:04.926982 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550830-n66j7" event={"ID":"4bac7da7-e0f6-4875-b683-9dd6842c915b","Type":"ContainerDied","Data":"d81de4dfe6952846acab2f7b1df19fb20feb5a135536a7478b7b6bfe7edc5207"} Mar 09 09:50:04 crc kubenswrapper[4692]: I0309 09:50:04.927030 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d81de4dfe6952846acab2f7b1df19fb20feb5a135536a7478b7b6bfe7edc5207" Mar 09 09:50:04 crc kubenswrapper[4692]: I0309 09:50:04.927089 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550830-n66j7" Mar 09 09:50:05 crc kubenswrapper[4692]: I0309 09:50:05.235178 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550824-cjjf5"] Mar 09 09:50:05 crc kubenswrapper[4692]: I0309 09:50:05.240667 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550824-cjjf5"] Mar 09 09:50:06 crc kubenswrapper[4692]: I0309 09:50:06.081365 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac8b5345-e2ae-419d-8de5-b36244b0cf42" path="/var/lib/kubelet/pods/ac8b5345-e2ae-419d-8de5-b36244b0cf42/volumes" Mar 09 09:50:13 crc kubenswrapper[4692]: I0309 09:50:13.071845 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:50:13 crc kubenswrapper[4692]: E0309 09:50:13.072609 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:50:24 crc kubenswrapper[4692]: I0309 09:50:24.071940 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:50:24 crc kubenswrapper[4692]: E0309 09:50:24.073423 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:50:37 crc kubenswrapper[4692]: I0309 09:50:37.071736 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:50:37 crc kubenswrapper[4692]: E0309 09:50:37.072557 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:50:51 crc kubenswrapper[4692]: I0309 09:50:51.072013 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:50:52 crc kubenswrapper[4692]: I0309 09:50:52.248710 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"cfb6f9265bcdee5d06e08425c1dd3a025aa1540a263740462908a83a3191a85f"} Mar 09 09:51:03 crc kubenswrapper[4692]: I0309 09:51:03.826038 4692 scope.go:117] "RemoveContainer" containerID="448c9f3423d1ab0e1159af44b656256ded9872b1c632b1fa393e35c24184d419" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.058947 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnnst"] Mar 09 09:51:10 crc kubenswrapper[4692]: E0309 09:51:10.060860 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bac7da7-e0f6-4875-b683-9dd6842c915b" containerName="oc" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.060882 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bac7da7-e0f6-4875-b683-9dd6842c915b" containerName="oc" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.061066 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bac7da7-e0f6-4875-b683-9dd6842c915b" containerName="oc" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.061655 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.065860 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.066215 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnnst"] Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.066272 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.168258 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-swiftconf\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.168324 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-dispersionconf\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.168399 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-ring-data-devices\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.168456 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfgvm\" (UniqueName: \"kubernetes.io/projected/7246e0dd-8bcf-47aa-a375-facc99397606-kube-api-access-vfgvm\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.168613 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-scripts\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.168729 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7246e0dd-8bcf-47aa-a375-facc99397606-etc-swift\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.270260 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-swiftconf\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.270310 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-dispersionconf\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.270335 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-ring-data-devices\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.270379 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfgvm\" (UniqueName: \"kubernetes.io/projected/7246e0dd-8bcf-47aa-a375-facc99397606-kube-api-access-vfgvm\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.270419 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-scripts\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.270439 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7246e0dd-8bcf-47aa-a375-facc99397606-etc-swift\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.271093 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7246e0dd-8bcf-47aa-a375-facc99397606-etc-swift\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.271282 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-ring-data-devices\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.271452 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-scripts\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.276505 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-dispersionconf\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.285589 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-swiftconf\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.290518 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfgvm\" (UniqueName: \"kubernetes.io/projected/7246e0dd-8bcf-47aa-a375-facc99397606-kube-api-access-vfgvm\") pod \"swift-ring-rebalance-debug-vnnst\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.406190 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:10 crc kubenswrapper[4692]: I0309 09:51:10.842311 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnnst"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.131516 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.137184 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.137552 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.142215 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.157840 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.191911 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm6hg\" (UniqueName: \"kubernetes.io/projected/e35fd62c-2bc9-44a1-9541-691741fb1d69-kube-api-access-dm6hg\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.192288 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e35fd62c-2bc9-44a1-9541-691741fb1d69-cache\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.192414 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.192439 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e35fd62c-2bc9-44a1-9541-691741fb1d69-etc-swift\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.192823 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e35fd62c-2bc9-44a1-9541-691741fb1d69-lock\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.211224 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.282402 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-knbvh"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.289010 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-knbvh"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294471 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm6hg\" (UniqueName: \"kubernetes.io/projected/e35fd62c-2bc9-44a1-9541-691741fb1d69-kube-api-access-dm6hg\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294667 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e35fd62c-2bc9-44a1-9541-691741fb1d69-cache\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294715 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st5mm\" (UniqueName: \"kubernetes.io/projected/367e8bc6-334a-490d-a1a5-4969f2a74a52-kube-api-access-st5mm\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294745 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294763 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e35fd62c-2bc9-44a1-9541-691741fb1d69-etc-swift\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294792 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/367e8bc6-334a-490d-a1a5-4969f2a74a52-lock\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294840 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294876 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e35fd62c-2bc9-44a1-9541-691741fb1d69-lock\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294918 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/367e8bc6-334a-490d-a1a5-4969f2a74a52-etc-swift\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.294940 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/367e8bc6-334a-490d-a1a5-4969f2a74a52-cache\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.295092 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") device mount path \"/mnt/openstack/pv11\"" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.295381 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e35fd62c-2bc9-44a1-9541-691741fb1d69-cache\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.296561 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e35fd62c-2bc9-44a1-9541-691741fb1d69-lock\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.303936 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e35fd62c-2bc9-44a1-9541-691741fb1d69-etc-swift\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.311883 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-p6vr9"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.312953 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.319696 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm6hg\" (UniqueName: \"kubernetes.io/projected/e35fd62c-2bc9-44a1-9541-691741fb1d69-kube-api-access-dm6hg\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.324453 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-p6vr9"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.347555 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-1\" (UID: \"e35fd62c-2bc9-44a1-9541-691741fb1d69\") " pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.374918 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" event={"ID":"7246e0dd-8bcf-47aa-a375-facc99397606","Type":"ContainerStarted","Data":"21a1e11b7f3d4f058b58cb33009cee8f857f530015b70a25acc6fb0d7936bcc6"} Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.374982 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" event={"ID":"7246e0dd-8bcf-47aa-a375-facc99397606","Type":"ContainerStarted","Data":"b17219bed0e3f988058751ff55e9ce09bbddf33935cc021cdb5e18f05d038f59"} Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396660 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st5mm\" (UniqueName: \"kubernetes.io/projected/367e8bc6-334a-490d-a1a5-4969f2a74a52-kube-api-access-st5mm\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396728 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-scripts\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396758 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/367e8bc6-334a-490d-a1a5-4969f2a74a52-lock\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396805 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-ring-data-devices\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396833 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396878 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvz26\" (UniqueName: \"kubernetes.io/projected/5a69287f-b725-45f6-85ed-9ce460d1e031-kube-api-access-mvz26\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396904 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-swiftconf\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396923 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-dispersionconf\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396948 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/367e8bc6-334a-490d-a1a5-4969f2a74a52-cache\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396968 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/367e8bc6-334a-490d-a1a5-4969f2a74a52-etc-swift\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.396997 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5a69287f-b725-45f6-85ed-9ce460d1e031-etc-swift\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.397665 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") device mount path \"/mnt/openstack/pv02\"" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.400859 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" podStartSLOduration=1.400840434 podStartE2EDuration="1.400840434s" podCreationTimestamp="2026-03-09 09:51:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:11.399489996 +0000 UTC m=+1872.224225587" watchObservedRunningTime="2026-03-09 09:51:11.400840434 +0000 UTC m=+1872.225576015" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.402522 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/367e8bc6-334a-490d-a1a5-4969f2a74a52-lock\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.402717 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/367e8bc6-334a-490d-a1a5-4969f2a74a52-cache\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.404225 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/367e8bc6-334a-490d-a1a5-4969f2a74a52-etc-swift\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.418423 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st5mm\" (UniqueName: \"kubernetes.io/projected/367e8bc6-334a-490d-a1a5-4969f2a74a52-kube-api-access-st5mm\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.420981 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-2\" (UID: \"367e8bc6-334a-490d-a1a5-4969f2a74a52\") " pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.498783 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-scripts\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.498880 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-ring-data-devices\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.498918 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvz26\" (UniqueName: \"kubernetes.io/projected/5a69287f-b725-45f6-85ed-9ce460d1e031-kube-api-access-mvz26\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.498938 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-dispersionconf\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.498951 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-swiftconf\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.498980 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5a69287f-b725-45f6-85ed-9ce460d1e031-etc-swift\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.501037 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-scripts\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.501668 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-ring-data-devices\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.502649 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-1" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.502758 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5a69287f-b725-45f6-85ed-9ce460d1e031-etc-swift\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.505871 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-dispersionconf\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.507399 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-swiftconf\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.510133 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-2" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.550655 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvz26\" (UniqueName: \"kubernetes.io/projected/5a69287f-b725-45f6-85ed-9ce460d1e031-kube-api-access-mvz26\") pod \"swift-ring-rebalance-p6vr9\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.557445 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-69ksf"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.560798 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.584652 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-69ksf"] Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.601725 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1028ab76-0253-4e6d-9c68-d0b4758a9182-log-httpd\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.602034 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1028ab76-0253-4e6d-9c68-d0b4758a9182-etc-swift\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.602698 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97sxx\" (UniqueName: \"kubernetes.io/projected/1028ab76-0253-4e6d-9c68-d0b4758a9182-kube-api-access-97sxx\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.602803 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1028ab76-0253-4e6d-9c68-d0b4758a9182-config-data\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.603044 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1028ab76-0253-4e6d-9c68-d0b4758a9182-run-httpd\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.702856 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.705244 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1028ab76-0253-4e6d-9c68-d0b4758a9182-run-httpd\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.705284 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1028ab76-0253-4e6d-9c68-d0b4758a9182-log-httpd\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.705318 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1028ab76-0253-4e6d-9c68-d0b4758a9182-etc-swift\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.705337 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97sxx\" (UniqueName: \"kubernetes.io/projected/1028ab76-0253-4e6d-9c68-d0b4758a9182-kube-api-access-97sxx\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.705357 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1028ab76-0253-4e6d-9c68-d0b4758a9182-config-data\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.707651 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1028ab76-0253-4e6d-9c68-d0b4758a9182-run-httpd\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.707963 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1028ab76-0253-4e6d-9c68-d0b4758a9182-log-httpd\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.716340 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1028ab76-0253-4e6d-9c68-d0b4758a9182-etc-swift\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.716985 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1028ab76-0253-4e6d-9c68-d0b4758a9182-config-data\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.724838 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97sxx\" (UniqueName: \"kubernetes.io/projected/1028ab76-0253-4e6d-9c68-d0b4758a9182-kube-api-access-97sxx\") pod \"swift-proxy-76c998454c-69ksf\" (UID: \"1028ab76-0253-4e6d-9c68-d0b4758a9182\") " pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:11 crc kubenswrapper[4692]: I0309 09:51:11.891911 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:12 crc kubenswrapper[4692]: W0309 09:51:12.025608 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode35fd62c_2bc9_44a1_9541_691741fb1d69.slice/crio-9b3defe4246fef101043e4c28c68aa1ad56ed0fd444fbaa810732b2da0ecf201 WatchSource:0}: Error finding container 9b3defe4246fef101043e4c28c68aa1ad56ed0fd444fbaa810732b2da0ecf201: Status 404 returned error can't find the container with id 9b3defe4246fef101043e4c28c68aa1ad56ed0fd444fbaa810732b2da0ecf201 Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.026490 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-1"] Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.082439 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70cc1105-fd62-42e9-b6cb-612adea92724" path="/var/lib/kubelet/pods/70cc1105-fd62-42e9-b6cb-612adea92724/volumes" Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.082958 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-2"] Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.138898 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-p6vr9"] Mar 09 09:51:12 crc kubenswrapper[4692]: W0309 09:51:12.142135 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a69287f_b725_45f6_85ed_9ce460d1e031.slice/crio-5c6f1805d18fe55a043063aa57af4ec327cd7276b529680b415ab5177369a219 WatchSource:0}: Error finding container 5c6f1805d18fe55a043063aa57af4ec327cd7276b529680b415ab5177369a219: Status 404 returned error can't find the container with id 5c6f1805d18fe55a043063aa57af4ec327cd7276b529680b415ab5177369a219 Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.341290 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-proxy-76c998454c-69ksf"] Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.455594 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" event={"ID":"5a69287f-b725-45f6-85ed-9ce460d1e031","Type":"ContainerStarted","Data":"15a64cc3f8b1ab07b26622c0a3443f9896d814d36269ba3d52e228e8d86083b6"} Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.455666 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" event={"ID":"5a69287f-b725-45f6-85ed-9ce460d1e031","Type":"ContainerStarted","Data":"5c6f1805d18fe55a043063aa57af4ec327cd7276b529680b415ab5177369a219"} Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.487546 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" podStartSLOduration=1.487530382 podStartE2EDuration="1.487530382s" podCreationTimestamp="2026-03-09 09:51:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:12.48463304 +0000 UTC m=+1873.309368631" watchObservedRunningTime="2026-03-09 09:51:12.487530382 +0000 UTC m=+1873.312265963" Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.489473 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" event={"ID":"1028ab76-0253-4e6d-9c68-d0b4758a9182","Type":"ContainerStarted","Data":"d1208cc5ab50f35139b841d86316e3065acc56e06f5d5d9f00914becf196b5c0"} Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.508261 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"dd3bc285dad58e1a543c821883eb3a4f1539b9f44500d4c4f6dea8c078a3042e"} Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.508317 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"a390cd29aa40f6a7eab7651fe380f5c3d7d736d4afaed2630e683171bd9ed133"} Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.520272 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"00068029a79342ae13cd81d86cd8e915f5d6ba9a7e160271a21d7810392a5df8"} Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.520308 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"3b18fcb2cd30477fc57f6edf64d7eb49c74133a9ac91d8108d13d4a819b3f7a9"} Mar 09 09:51:12 crc kubenswrapper[4692]: I0309 09:51:12.520319 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"9b3defe4246fef101043e4c28c68aa1ad56ed0fd444fbaa810732b2da0ecf201"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.530092 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" event={"ID":"1028ab76-0253-4e6d-9c68-d0b4758a9182","Type":"ContainerStarted","Data":"a4b3cdcf86f06d40e399121d8aedc72d6530defcca99fbe80c2aab922e672e44"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.531368 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" event={"ID":"1028ab76-0253-4e6d-9c68-d0b4758a9182","Type":"ContainerStarted","Data":"7f7afae8b15660fefb13d06c97a9fda08451265f703658561f85f2fe4415a560"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.531483 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.531589 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.535946 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"9213932a23740010af923fdfca4435e12eef4b50ee8e4d28dc5f7206c3c1b279"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.536084 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"9fe5532b08d3b0d34c6538e7957140341ef8afdd93ce340deddf0b0df711cbe4"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.536203 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"faa71d6f4c989299711cc96826bfe848afe5370d99efc20506bc3106ac9a4005"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.536300 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"6461987093c5d33964ef258daffe39246de235fb1fd7f1ecf0e4ad779506fd66"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.536446 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"c0332bfebb8b0a3b05e87ad22ce5c945f21095e10fd7955df96674bbde067c39"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.543881 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"c40dd3f6eadf66fe0400281613f737d04f9086cd94b58597a33e12c13b978b65"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.544113 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"de485c649ed59ee939a6e1f7be954a261a48e0f4c1187dc9bb5f12e7ef440e3e"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.544194 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"648bb95ee4a386d990e44c32d804f4939a10ede4889361661cb35257cde8f82f"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.544322 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"5e2d42fc38b81c5285e89f19dc55500219b9bfe5b4155afd3a8b50936af6ab75"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.544387 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"8bf6bcfa818210100b54c1852a6e4971db7220d3dd01b4740876346425925303"} Mar 09 09:51:13 crc kubenswrapper[4692]: I0309 09:51:13.564272 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" podStartSLOduration=2.564252789 podStartE2EDuration="2.564252789s" podCreationTimestamp="2026-03-09 09:51:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:13.559878595 +0000 UTC m=+1874.384614196" watchObservedRunningTime="2026-03-09 09:51:13.564252789 +0000 UTC m=+1874.388988370" Mar 09 09:51:14 crc kubenswrapper[4692]: I0309 09:51:14.630216 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"74d9a351db7a08edf389ebd7326b505198200ffcb7a73c9e56b4aba4c3732d9e"} Mar 09 09:51:14 crc kubenswrapper[4692]: I0309 09:51:14.634046 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"0c2f38157fb3081732a3f7c2eb8a888047da495242b0934fba34a8c86c1c1d0b"} Mar 09 09:51:14 crc kubenswrapper[4692]: I0309 09:51:14.634102 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"42904dec2bc30ae287177e6105c39571cf3f31b6d02c1e870e947168aff575e9"} Mar 09 09:51:14 crc kubenswrapper[4692]: I0309 09:51:14.635908 4692 generic.go:334] "Generic (PLEG): container finished" podID="7246e0dd-8bcf-47aa-a375-facc99397606" containerID="21a1e11b7f3d4f058b58cb33009cee8f857f530015b70a25acc6fb0d7936bcc6" exitCode=0 Mar 09 09:51:14 crc kubenswrapper[4692]: I0309 09:51:14.636658 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" event={"ID":"7246e0dd-8bcf-47aa-a375-facc99397606","Type":"ContainerDied","Data":"21a1e11b7f3d4f058b58cb33009cee8f857f530015b70a25acc6fb0d7936bcc6"} Mar 09 09:51:16 crc kubenswrapper[4692]: I0309 09:51:16.198799 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"1b5e01f8bc051daf97212585c15d150659bd1ff2058d232786c9ad19c7dc3607"} Mar 09 09:51:16 crc kubenswrapper[4692]: I0309 09:51:16.215664 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"ca91ac42f3df9c02abffd8a7f1eff488bf0b45068cabb6f8e8992c5618493a67"} Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.232759 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pfk8k"] Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.238235 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.291182 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pfk8k"] Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.392329 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt974\" (UniqueName: \"kubernetes.io/projected/86ec167e-b651-4c2b-87a4-d853a9340409-kube-api-access-tt974\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.392376 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-catalog-content\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.392414 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-utilities\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.571127 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt974\" (UniqueName: \"kubernetes.io/projected/86ec167e-b651-4c2b-87a4-d853a9340409-kube-api-access-tt974\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.571531 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-catalog-content\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.571571 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-utilities\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.578657 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-catalog-content\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.578872 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-utilities\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.596506 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"57540566ab41847a7c356a8837b77f8e4b1579f670a06259a2140885f97e46eb"} Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.601370 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"5ed16746da149872f3f57b7b8f4c0678ffc4d8da962cdcd44933c2e7203803d2"} Mar 09 09:51:17 crc kubenswrapper[4692]: I0309 09:51:17.639464 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt974\" (UniqueName: \"kubernetes.io/projected/86ec167e-b651-4c2b-87a4-d853a9340409-kube-api-access-tt974\") pod \"certified-operators-pfk8k\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:18 crc kubenswrapper[4692]: I0309 09:51:18.166062 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:18 crc kubenswrapper[4692]: I0309 09:51:18.826924 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"da8c990ad0b61d83320566ee7406f657e3d32debca0c07df09887b0ae5da9af8"} Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.207748 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.255734 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnnst"] Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.270727 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vnnst"] Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.296723 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-scripts\") pod \"7246e0dd-8bcf-47aa-a375-facc99397606\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.296857 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7246e0dd-8bcf-47aa-a375-facc99397606-etc-swift\") pod \"7246e0dd-8bcf-47aa-a375-facc99397606\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.296999 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfgvm\" (UniqueName: \"kubernetes.io/projected/7246e0dd-8bcf-47aa-a375-facc99397606-kube-api-access-vfgvm\") pod \"7246e0dd-8bcf-47aa-a375-facc99397606\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.297045 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-dispersionconf\") pod \"7246e0dd-8bcf-47aa-a375-facc99397606\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.297099 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-swiftconf\") pod \"7246e0dd-8bcf-47aa-a375-facc99397606\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.297134 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-ring-data-devices\") pod \"7246e0dd-8bcf-47aa-a375-facc99397606\" (UID: \"7246e0dd-8bcf-47aa-a375-facc99397606\") " Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.298273 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7246e0dd-8bcf-47aa-a375-facc99397606" (UID: "7246e0dd-8bcf-47aa-a375-facc99397606"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.299035 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7246e0dd-8bcf-47aa-a375-facc99397606-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7246e0dd-8bcf-47aa-a375-facc99397606" (UID: "7246e0dd-8bcf-47aa-a375-facc99397606"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.328528 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7246e0dd-8bcf-47aa-a375-facc99397606-kube-api-access-vfgvm" (OuterVolumeSpecName: "kube-api-access-vfgvm") pod "7246e0dd-8bcf-47aa-a375-facc99397606" (UID: "7246e0dd-8bcf-47aa-a375-facc99397606"). InnerVolumeSpecName "kube-api-access-vfgvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.359687 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7246e0dd-8bcf-47aa-a375-facc99397606" (UID: "7246e0dd-8bcf-47aa-a375-facc99397606"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.400613 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.400669 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfgvm\" (UniqueName: \"kubernetes.io/projected/7246e0dd-8bcf-47aa-a375-facc99397606-kube-api-access-vfgvm\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.400685 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.400699 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7246e0dd-8bcf-47aa-a375-facc99397606-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.415810 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7246e0dd-8bcf-47aa-a375-facc99397606" (UID: "7246e0dd-8bcf-47aa-a375-facc99397606"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.431640 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx"] Mar 09 09:51:19 crc kubenswrapper[4692]: E0309 09:51:19.432010 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7246e0dd-8bcf-47aa-a375-facc99397606" containerName="swift-ring-rebalance" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.432036 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="7246e0dd-8bcf-47aa-a375-facc99397606" containerName="swift-ring-rebalance" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.432297 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="7246e0dd-8bcf-47aa-a375-facc99397606" containerName="swift-ring-rebalance" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.432927 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.437348 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-scripts" (OuterVolumeSpecName: "scripts") pod "7246e0dd-8bcf-47aa-a375-facc99397606" (UID: "7246e0dd-8bcf-47aa-a375-facc99397606"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.492933 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx"] Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504367 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd29f9fc-f6f2-4520-8826-861fa4166a70-etc-swift\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504462 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-dispersionconf\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504509 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56pf8\" (UniqueName: \"kubernetes.io/projected/bd29f9fc-f6f2-4520-8826-861fa4166a70-kube-api-access-56pf8\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504544 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-scripts\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504583 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-ring-data-devices\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504608 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504666 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7246e0dd-8bcf-47aa-a375-facc99397606-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.504679 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7246e0dd-8bcf-47aa-a375-facc99397606-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.605609 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-dispersionconf\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.605685 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56pf8\" (UniqueName: \"kubernetes.io/projected/bd29f9fc-f6f2-4520-8826-861fa4166a70-kube-api-access-56pf8\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.605727 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-scripts\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.605768 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-ring-data-devices\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.605795 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.605832 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd29f9fc-f6f2-4520-8826-861fa4166a70-etc-swift\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.606366 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd29f9fc-f6f2-4520-8826-861fa4166a70-etc-swift\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.607023 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-scripts\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.607549 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-ring-data-devices\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.698314 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56pf8\" (UniqueName: \"kubernetes.io/projected/bd29f9fc-f6f2-4520-8826-861fa4166a70-kube-api-access-56pf8\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.698677 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-dispersionconf\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.706407 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf\") pod \"swift-ring-rebalance-debug-8xrqx\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:19 crc kubenswrapper[4692]: I0309 09:51:19.919763 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.025716 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b17219bed0e3f988058751ff55e9ce09bbddf33935cc021cdb5e18f05d038f59" Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.025810 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vnnst" Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.163478 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7246e0dd-8bcf-47aa-a375-facc99397606" path="/var/lib/kubelet/pods/7246e0dd-8bcf-47aa-a375-facc99397606/volumes" Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.414475 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"806f2f40f37068851e222107c30e858b0054f8467a7698a915877f4a610b508b"} Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.425089 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pfk8k"] Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.479259 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"3477e5b23371874057c8bdda28e83054f9c14bbc3a85b8dc1387233ce4cd164c"} Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.479299 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"584b931ccfa381dc00e890511bff00218f1fd05c8bd1be219e7be6fcffbffbe6"} Mar 09 09:51:20 crc kubenswrapper[4692]: I0309 09:51:20.640079 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx"] Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.455308 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx"] Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.518901 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfk8k" event={"ID":"86ec167e-b651-4c2b-87a4-d853a9340409","Type":"ContainerStarted","Data":"ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86"} Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.518973 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfk8k" event={"ID":"86ec167e-b651-4c2b-87a4-d853a9340409","Type":"ContainerStarted","Data":"bfc8fedbcd172b2a7a20575735324b2debfa2af27ee8d748bc01bbb382832b79"} Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.677499 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"98287cbf35909a50c94a730ef576704fedd21a70472fa1deb67894125909d65d"} Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.704102 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"3f9549f351cc1b6d799d0a198b4ad3d28c3dbfecbba263ad60d5c4fe455d6d1d"} Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.729557 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" event={"ID":"bd29f9fc-f6f2-4520-8826-861fa4166a70","Type":"ContainerStarted","Data":"4d5d37820e3434f572f1f76434ca3a90e3138be731b570755a7493e1d1b1e9d2"} Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.930298 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:21 crc kubenswrapper[4692]: I0309 09:51:21.988245 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="swift-kuttl-tests/swift-proxy-76c998454c-69ksf" Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.109190 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-bfcdc9759-4998p"] Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.109455 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-httpd" containerID="cri-o://bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.109597 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-server" containerID="cri-o://95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.745080 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-1" event={"ID":"e35fd62c-2bc9-44a1-9541-691741fb1d69","Type":"ContainerStarted","Data":"c57cce64d089e16fb43eaf5d8f6c1dabd2995cfba6bc2d2f03844e5f4284ed18"} Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.747969 4692 generic.go:334] "Generic (PLEG): container finished" podID="bd23df0a-a16c-4a17-8308-909043e28377" containerID="bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9" exitCode=0 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.748071 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" event={"ID":"bd23df0a-a16c-4a17-8308-909043e28377","Type":"ContainerDied","Data":"bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9"} Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.755105 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"2512f5b44e92307b7f6179c48e65282fda46d1ded0c505bec75cf377c7457fd7"} Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.755192 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"44209ba7f8e5864a519425bac80ee74fac8372a3810f668efdc17536aee859b0"} Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.755205 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-2" event={"ID":"367e8bc6-334a-490d-a1a5-4969f2a74a52","Type":"ContainerStarted","Data":"ce0b84abf71101b5b162b723668b35febf5694c6839951a38841a8d52d4c538b"} Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.756822 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" event={"ID":"bd29f9fc-f6f2-4520-8826-861fa4166a70","Type":"ContainerStarted","Data":"2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c"} Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.756881 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" podUID="bd29f9fc-f6f2-4520-8826-861fa4166a70" containerName="swift-ring-rebalance" containerID="cri-o://2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.758863 4692 generic.go:334] "Generic (PLEG): container finished" podID="86ec167e-b651-4c2b-87a4-d853a9340409" containerID="ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86" exitCode=0 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.758902 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfk8k" event={"ID":"86ec167e-b651-4c2b-87a4-d853a9340409","Type":"ContainerDied","Data":"ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86"} Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.786112 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-1" podStartSLOduration=12.786093527 podStartE2EDuration="12.786093527s" podCreationTimestamp="2026-03-09 09:51:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:22.782713792 +0000 UTC m=+1883.607449383" watchObservedRunningTime="2026-03-09 09:51:22.786093527 +0000 UTC m=+1883.610829128" Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.839284 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" podStartSLOduration=3.8392694609999998 podStartE2EDuration="3.839269461s" podCreationTimestamp="2026-03-09 09:51:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:22.835425323 +0000 UTC m=+1883.660160924" watchObservedRunningTime="2026-03-09 09:51:22.839269461 +0000 UTC m=+1883.664005042" Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.877981 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-2" podStartSLOduration=12.877963816 podStartE2EDuration="12.877963816s" podCreationTimestamp="2026-03-09 09:51:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:22.874156238 +0000 UTC m=+1883.698891819" watchObservedRunningTime="2026-03-09 09:51:22.877963816 +0000 UTC m=+1883.702699397" Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900102 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900705 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-server" containerID="cri-o://ab65f21b397d18fb36d0d4050eebb38914e73ef3e17f18703e7a399455036458" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900750 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-server" containerID="cri-o://e93f97a75cd47e8689825ca7c8a46036015ce8c783fbd946e4b5aae4c8f79b03" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900850 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-updater" containerID="cri-o://9c329076d4b5930b205901b8cdceb024a0470b51a44c26609e18b30b6e9ff6b9" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900890 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-sharder" containerID="cri-o://58b3d88d99997185e1fe0c65177b1557bbf0f1600dfcd4c1a78a97dfe4cc3869" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900902 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-auditor" containerID="cri-o://4b04d6a9a6046a7cf5adeb6b3544acc38445e80b0d6ec6bd670e88aac9bee508" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900943 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="swift-recon-cron" containerID="cri-o://ab12c78b3d286b45efb3c2c57c2505919e7441ae58bbc75580420db1252bd058" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900959 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-server" containerID="cri-o://3c66beccc4342856159e6f72d1d1ae309389f4f7b64191912f7f8ea930ad529f" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.900946 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-replicator" containerID="cri-o://2ea0a775295e7a2ebd710a249b8eef9e502086a8adca32469d0a7e5f0360946b" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901001 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="rsync" containerID="cri-o://2394ec11824f008a37542215a330958d91bdf6a7d65a1fd73ce74901131426d5" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901028 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-reaper" containerID="cri-o://c280851b4e3b0f461e368eb3548350b7ce5623cbea4928739b6a1c9c9b6e4b82" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901044 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-expirer" containerID="cri-o://703b6e2cbdc86106d9e4b0e1d0f94ab5534c785c45615232e8488b6c7384526e" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901069 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-auditor" containerID="cri-o://bb413cb17c31ff233ceedc8a35e86e5b01a697b0c889b4a43599446ef0c3e78d" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901078 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-updater" containerID="cri-o://564d437ed65b90cb81f73d2a130ee67b0cedeedb41b6b4d9a24f4f61cfa50bcd" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901111 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-replicator" containerID="cri-o://c4ecffd2dfe8323c86c3932e7f158bba159aef48ce3a7407fa6ec794dcb21fb9" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901130 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-replicator" containerID="cri-o://d96fdc6db8ea6a8d4119dbbc1739cb93789eee6ec1375ebdcdbbcaac8fce23fb" gracePeriod=30 Mar 09 09:51:22 crc kubenswrapper[4692]: I0309 09:51:22.901117 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="swift-kuttl-tests/swift-storage-0" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-auditor" containerID="cri-o://06be2bd359a325cfb2ef0967a7db004ec3b7c9adb30c68c9aa995111081731ee" gracePeriod=30 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.373952 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.495692 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-etc-swift\") pod \"bd23df0a-a16c-4a17-8308-909043e28377\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.495740 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-run-httpd\") pod \"bd23df0a-a16c-4a17-8308-909043e28377\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.495867 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23df0a-a16c-4a17-8308-909043e28377-config-data\") pod \"bd23df0a-a16c-4a17-8308-909043e28377\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.495957 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-log-httpd\") pod \"bd23df0a-a16c-4a17-8308-909043e28377\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.495985 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7cf6\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-kube-api-access-z7cf6\") pod \"bd23df0a-a16c-4a17-8308-909043e28377\" (UID: \"bd23df0a-a16c-4a17-8308-909043e28377\") " Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.496370 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bd23df0a-a16c-4a17-8308-909043e28377" (UID: "bd23df0a-a16c-4a17-8308-909043e28377"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.496664 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bd23df0a-a16c-4a17-8308-909043e28377" (UID: "bd23df0a-a16c-4a17-8308-909043e28377"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.502005 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-kube-api-access-z7cf6" (OuterVolumeSpecName: "kube-api-access-z7cf6") pod "bd23df0a-a16c-4a17-8308-909043e28377" (UID: "bd23df0a-a16c-4a17-8308-909043e28377"). InnerVolumeSpecName "kube-api-access-z7cf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.502204 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bd23df0a-a16c-4a17-8308-909043e28377" (UID: "bd23df0a-a16c-4a17-8308-909043e28377"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.535497 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd23df0a-a16c-4a17-8308-909043e28377-config-data" (OuterVolumeSpecName: "config-data") pod "bd23df0a-a16c-4a17-8308-909043e28377" (UID: "bd23df0a-a16c-4a17-8308-909043e28377"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.597921 4692 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd23df0a-a16c-4a17-8308-909043e28377-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.597963 4692 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.597978 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7cf6\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-kube-api-access-z7cf6\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.597993 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/bd23df0a-a16c-4a17-8308-909043e28377-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.598001 4692 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd23df0a-a16c-4a17-8308-909043e28377-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.796504 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="58b3d88d99997185e1fe0c65177b1557bbf0f1600dfcd4c1a78a97dfe4cc3869" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.797644 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="2394ec11824f008a37542215a330958d91bdf6a7d65a1fd73ce74901131426d5" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.797742 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="703b6e2cbdc86106d9e4b0e1d0f94ab5534c785c45615232e8488b6c7384526e" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.797852 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="564d437ed65b90cb81f73d2a130ee67b0cedeedb41b6b4d9a24f4f61cfa50bcd" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.797936 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="06be2bd359a325cfb2ef0967a7db004ec3b7c9adb30c68c9aa995111081731ee" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798035 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="d96fdc6db8ea6a8d4119dbbc1739cb93789eee6ec1375ebdcdbbcaac8fce23fb" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798127 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="e93f97a75cd47e8689825ca7c8a46036015ce8c783fbd946e4b5aae4c8f79b03" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798245 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="9c329076d4b5930b205901b8cdceb024a0470b51a44c26609e18b30b6e9ff6b9" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.796586 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"58b3d88d99997185e1fe0c65177b1557bbf0f1600dfcd4c1a78a97dfe4cc3869"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798355 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"2394ec11824f008a37542215a330958d91bdf6a7d65a1fd73ce74901131426d5"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798375 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"703b6e2cbdc86106d9e4b0e1d0f94ab5534c785c45615232e8488b6c7384526e"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798385 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"564d437ed65b90cb81f73d2a130ee67b0cedeedb41b6b4d9a24f4f61cfa50bcd"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798394 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"06be2bd359a325cfb2ef0967a7db004ec3b7c9adb30c68c9aa995111081731ee"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798402 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"d96fdc6db8ea6a8d4119dbbc1739cb93789eee6ec1375ebdcdbbcaac8fce23fb"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798410 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"e93f97a75cd47e8689825ca7c8a46036015ce8c783fbd946e4b5aae4c8f79b03"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798419 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"9c329076d4b5930b205901b8cdceb024a0470b51a44c26609e18b30b6e9ff6b9"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798317 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="4b04d6a9a6046a7cf5adeb6b3544acc38445e80b0d6ec6bd670e88aac9bee508" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798462 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="2ea0a775295e7a2ebd710a249b8eef9e502086a8adca32469d0a7e5f0360946b" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798483 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="3c66beccc4342856159e6f72d1d1ae309389f4f7b64191912f7f8ea930ad529f" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798494 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="c280851b4e3b0f461e368eb3548350b7ce5623cbea4928739b6a1c9c9b6e4b82" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798505 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="bb413cb17c31ff233ceedc8a35e86e5b01a697b0c889b4a43599446ef0c3e78d" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798517 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="c4ecffd2dfe8323c86c3932e7f158bba159aef48ce3a7407fa6ec794dcb21fb9" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798526 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="ab65f21b397d18fb36d0d4050eebb38914e73ef3e17f18703e7a399455036458" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798427 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"4b04d6a9a6046a7cf5adeb6b3544acc38445e80b0d6ec6bd670e88aac9bee508"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798616 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"2ea0a775295e7a2ebd710a249b8eef9e502086a8adca32469d0a7e5f0360946b"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798633 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"3c66beccc4342856159e6f72d1d1ae309389f4f7b64191912f7f8ea930ad529f"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798646 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"c280851b4e3b0f461e368eb3548350b7ce5623cbea4928739b6a1c9c9b6e4b82"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798660 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"bb413cb17c31ff233ceedc8a35e86e5b01a697b0c889b4a43599446ef0c3e78d"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798671 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"c4ecffd2dfe8323c86c3932e7f158bba159aef48ce3a7407fa6ec794dcb21fb9"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.798682 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"ab65f21b397d18fb36d0d4050eebb38914e73ef3e17f18703e7a399455036458"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.801305 4692 generic.go:334] "Generic (PLEG): container finished" podID="bd23df0a-a16c-4a17-8308-909043e28377" containerID="95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83" exitCode=0 Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.801409 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" event={"ID":"bd23df0a-a16c-4a17-8308-909043e28377","Type":"ContainerDied","Data":"95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.801454 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" event={"ID":"bd23df0a-a16c-4a17-8308-909043e28377","Type":"ContainerDied","Data":"e6766a81cb2b796bea4981966dc1d166f9eebd537b4e8e93480e92baf4238e1a"} Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.801477 4692 scope.go:117] "RemoveContainer" containerID="95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.801644 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-proxy-bfcdc9759-4998p" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.837768 4692 scope.go:117] "RemoveContainer" containerID="bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.845234 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-proxy-bfcdc9759-4998p"] Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.854036 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-proxy-bfcdc9759-4998p"] Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.856461 4692 scope.go:117] "RemoveContainer" containerID="95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83" Mar 09 09:51:23 crc kubenswrapper[4692]: E0309 09:51:23.857001 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83\": container with ID starting with 95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83 not found: ID does not exist" containerID="95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.857088 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83"} err="failed to get container status \"95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83\": rpc error: code = NotFound desc = could not find container \"95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83\": container with ID starting with 95bdf0c2d74b58f256c0fef23faf205a1dad07ab4835212bfafa61f95bfd3b83 not found: ID does not exist" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.857153 4692 scope.go:117] "RemoveContainer" containerID="bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9" Mar 09 09:51:23 crc kubenswrapper[4692]: E0309 09:51:23.857660 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9\": container with ID starting with bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9 not found: ID does not exist" containerID="bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9" Mar 09 09:51:23 crc kubenswrapper[4692]: I0309 09:51:23.857697 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9"} err="failed to get container status \"bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9\": rpc error: code = NotFound desc = could not find container \"bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9\": container with ID starting with bad270e8c3361fa1cfa6eb3edf44c4f271759815da06dd4bc6624a9e24a93ad9 not found: ID does not exist" Mar 09 09:51:24 crc kubenswrapper[4692]: I0309 09:51:24.080623 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23df0a-a16c-4a17-8308-909043e28377" path="/var/lib/kubelet/pods/bd23df0a-a16c-4a17-8308-909043e28377/volumes" Mar 09 09:51:24 crc kubenswrapper[4692]: I0309 09:51:24.812755 4692 generic.go:334] "Generic (PLEG): container finished" podID="86ec167e-b651-4c2b-87a4-d853a9340409" containerID="d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8" exitCode=0 Mar 09 09:51:24 crc kubenswrapper[4692]: I0309 09:51:24.812803 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfk8k" event={"ID":"86ec167e-b651-4c2b-87a4-d853a9340409","Type":"ContainerDied","Data":"d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8"} Mar 09 09:51:25 crc kubenswrapper[4692]: I0309 09:51:25.822778 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfk8k" event={"ID":"86ec167e-b651-4c2b-87a4-d853a9340409","Type":"ContainerStarted","Data":"dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49"} Mar 09 09:51:25 crc kubenswrapper[4692]: I0309 09:51:25.852462 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pfk8k" podStartSLOduration=6.380172462 podStartE2EDuration="8.852444062s" podCreationTimestamp="2026-03-09 09:51:17 +0000 UTC" firstStartedPulling="2026-03-09 09:51:22.760493133 +0000 UTC m=+1883.585228724" lastFinishedPulling="2026-03-09 09:51:25.232764743 +0000 UTC m=+1886.057500324" observedRunningTime="2026-03-09 09:51:25.84352242 +0000 UTC m=+1886.668258021" watchObservedRunningTime="2026-03-09 09:51:25.852444062 +0000 UTC m=+1886.677179633" Mar 09 09:51:28 crc kubenswrapper[4692]: I0309 09:51:28.166772 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:28 crc kubenswrapper[4692]: I0309 09:51:28.166832 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:28 crc kubenswrapper[4692]: I0309 09:51:28.213097 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:29 crc kubenswrapper[4692]: I0309 09:51:29.852905 4692 generic.go:334] "Generic (PLEG): container finished" podID="5a69287f-b725-45f6-85ed-9ce460d1e031" containerID="15a64cc3f8b1ab07b26622c0a3443f9896d814d36269ba3d52e228e8d86083b6" exitCode=0 Mar 09 09:51:29 crc kubenswrapper[4692]: I0309 09:51:29.852956 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" event={"ID":"5a69287f-b725-45f6-85ed-9ce460d1e031","Type":"ContainerDied","Data":"15a64cc3f8b1ab07b26622c0a3443f9896d814d36269ba3d52e228e8d86083b6"} Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.483614 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.634997 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-scripts\") pod \"5a69287f-b725-45f6-85ed-9ce460d1e031\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.635513 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-dispersionconf\") pod \"5a69287f-b725-45f6-85ed-9ce460d1e031\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.635539 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-ring-data-devices\") pod \"5a69287f-b725-45f6-85ed-9ce460d1e031\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.635596 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvz26\" (UniqueName: \"kubernetes.io/projected/5a69287f-b725-45f6-85ed-9ce460d1e031-kube-api-access-mvz26\") pod \"5a69287f-b725-45f6-85ed-9ce460d1e031\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.635674 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5a69287f-b725-45f6-85ed-9ce460d1e031-etc-swift\") pod \"5a69287f-b725-45f6-85ed-9ce460d1e031\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.635725 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-swiftconf\") pod \"5a69287f-b725-45f6-85ed-9ce460d1e031\" (UID: \"5a69287f-b725-45f6-85ed-9ce460d1e031\") " Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.636419 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5a69287f-b725-45f6-85ed-9ce460d1e031" (UID: "5a69287f-b725-45f6-85ed-9ce460d1e031"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.636574 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a69287f-b725-45f6-85ed-9ce460d1e031-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5a69287f-b725-45f6-85ed-9ce460d1e031" (UID: "5a69287f-b725-45f6-85ed-9ce460d1e031"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.643269 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a69287f-b725-45f6-85ed-9ce460d1e031-kube-api-access-mvz26" (OuterVolumeSpecName: "kube-api-access-mvz26") pod "5a69287f-b725-45f6-85ed-9ce460d1e031" (UID: "5a69287f-b725-45f6-85ed-9ce460d1e031"). InnerVolumeSpecName "kube-api-access-mvz26". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.655287 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-scripts" (OuterVolumeSpecName: "scripts") pod "5a69287f-b725-45f6-85ed-9ce460d1e031" (UID: "5a69287f-b725-45f6-85ed-9ce460d1e031"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.658001 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5a69287f-b725-45f6-85ed-9ce460d1e031" (UID: "5a69287f-b725-45f6-85ed-9ce460d1e031"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.659418 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5a69287f-b725-45f6-85ed-9ce460d1e031" (UID: "5a69287f-b725-45f6-85ed-9ce460d1e031"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.737007 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.737064 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.737078 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5a69287f-b725-45f6-85ed-9ce460d1e031-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.737100 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5a69287f-b725-45f6-85ed-9ce460d1e031-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.737115 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvz26\" (UniqueName: \"kubernetes.io/projected/5a69287f-b725-45f6-85ed-9ce460d1e031-kube-api-access-mvz26\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.737127 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5a69287f-b725-45f6-85ed-9ce460d1e031-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.872130 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" event={"ID":"5a69287f-b725-45f6-85ed-9ce460d1e031","Type":"ContainerDied","Data":"5c6f1805d18fe55a043063aa57af4ec327cd7276b529680b415ab5177369a219"} Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.872207 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c6f1805d18fe55a043063aa57af4ec327cd7276b529680b415ab5177369a219" Mar 09 09:51:31 crc kubenswrapper[4692]: I0309 09:51:31.872287 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-p6vr9" Mar 09 09:51:38 crc kubenswrapper[4692]: I0309 09:51:38.219831 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:38 crc kubenswrapper[4692]: I0309 09:51:38.281869 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pfk8k"] Mar 09 09:51:38 crc kubenswrapper[4692]: I0309 09:51:38.935070 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pfk8k" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="registry-server" containerID="cri-o://dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49" gracePeriod=2 Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.349560 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.456397 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt974\" (UniqueName: \"kubernetes.io/projected/86ec167e-b651-4c2b-87a4-d853a9340409-kube-api-access-tt974\") pod \"86ec167e-b651-4c2b-87a4-d853a9340409\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.456548 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-catalog-content\") pod \"86ec167e-b651-4c2b-87a4-d853a9340409\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.456580 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-utilities\") pod \"86ec167e-b651-4c2b-87a4-d853a9340409\" (UID: \"86ec167e-b651-4c2b-87a4-d853a9340409\") " Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.457481 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-utilities" (OuterVolumeSpecName: "utilities") pod "86ec167e-b651-4c2b-87a4-d853a9340409" (UID: "86ec167e-b651-4c2b-87a4-d853a9340409"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.462406 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86ec167e-b651-4c2b-87a4-d853a9340409-kube-api-access-tt974" (OuterVolumeSpecName: "kube-api-access-tt974") pod "86ec167e-b651-4c2b-87a4-d853a9340409" (UID: "86ec167e-b651-4c2b-87a4-d853a9340409"). InnerVolumeSpecName "kube-api-access-tt974". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.513349 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86ec167e-b651-4c2b-87a4-d853a9340409" (UID: "86ec167e-b651-4c2b-87a4-d853a9340409"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.558309 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.558346 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt974\" (UniqueName: \"kubernetes.io/projected/86ec167e-b651-4c2b-87a4-d853a9340409-kube-api-access-tt974\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.558359 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86ec167e-b651-4c2b-87a4-d853a9340409-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.946044 4692 generic.go:334] "Generic (PLEG): container finished" podID="86ec167e-b651-4c2b-87a4-d853a9340409" containerID="dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49" exitCode=0 Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.946121 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfk8k" event={"ID":"86ec167e-b651-4c2b-87a4-d853a9340409","Type":"ContainerDied","Data":"dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49"} Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.949074 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pfk8k" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.949152 4692 scope.go:117] "RemoveContainer" containerID="dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49" Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.946157 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pfk8k" event={"ID":"86ec167e-b651-4c2b-87a4-d853a9340409","Type":"ContainerDied","Data":"bfc8fedbcd172b2a7a20575735324b2debfa2af27ee8d748bc01bbb382832b79"} Mar 09 09:51:39 crc kubenswrapper[4692]: I0309 09:51:39.983649 4692 scope.go:117] "RemoveContainer" containerID="d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.000440 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pfk8k"] Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.008341 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pfk8k"] Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.016955 4692 scope.go:117] "RemoveContainer" containerID="ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.033754 4692 scope.go:117] "RemoveContainer" containerID="dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49" Mar 09 09:51:40 crc kubenswrapper[4692]: E0309 09:51:40.034263 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49\": container with ID starting with dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49 not found: ID does not exist" containerID="dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.034321 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49"} err="failed to get container status \"dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49\": rpc error: code = NotFound desc = could not find container \"dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49\": container with ID starting with dd9c7940bd163ff009ec4cc6eab7164809364100b7e2c108d0ba1a5caab22a49 not found: ID does not exist" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.034348 4692 scope.go:117] "RemoveContainer" containerID="d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8" Mar 09 09:51:40 crc kubenswrapper[4692]: E0309 09:51:40.035622 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8\": container with ID starting with d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8 not found: ID does not exist" containerID="d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.035653 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8"} err="failed to get container status \"d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8\": rpc error: code = NotFound desc = could not find container \"d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8\": container with ID starting with d1f9250f61fdfacd0cc3d250e8cf85c6eab467674855ef143a808b42970efeb8 not found: ID does not exist" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.035668 4692 scope.go:117] "RemoveContainer" containerID="ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86" Mar 09 09:51:40 crc kubenswrapper[4692]: E0309 09:51:40.035916 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86\": container with ID starting with ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86 not found: ID does not exist" containerID="ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.035936 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86"} err="failed to get container status \"ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86\": rpc error: code = NotFound desc = could not find container \"ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86\": container with ID starting with ee8f5f967a98c98953f2f3a93167a995c9de1562a0d5a89307872e5bf5ebbb86 not found: ID does not exist" Mar 09 09:51:40 crc kubenswrapper[4692]: I0309 09:51:40.080778 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" path="/var/lib/kubelet/pods/86ec167e-b651-4c2b-87a4-d853a9340409/volumes" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.651130 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.810824 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-scripts\") pod \"bd29f9fc-f6f2-4520-8826-861fa4166a70\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.810882 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf\") pod \"bd29f9fc-f6f2-4520-8826-861fa4166a70\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.810971 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56pf8\" (UniqueName: \"kubernetes.io/projected/bd29f9fc-f6f2-4520-8826-861fa4166a70-kube-api-access-56pf8\") pod \"bd29f9fc-f6f2-4520-8826-861fa4166a70\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.811024 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd29f9fc-f6f2-4520-8826-861fa4166a70-etc-swift\") pod \"bd29f9fc-f6f2-4520-8826-861fa4166a70\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.811099 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-dispersionconf\") pod \"bd29f9fc-f6f2-4520-8826-861fa4166a70\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.811124 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-ring-data-devices\") pod \"bd29f9fc-f6f2-4520-8826-861fa4166a70\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.812114 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bd29f9fc-f6f2-4520-8826-861fa4166a70" (UID: "bd29f9fc-f6f2-4520-8826-861fa4166a70"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.812175 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd29f9fc-f6f2-4520-8826-861fa4166a70-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bd29f9fc-f6f2-4520-8826-861fa4166a70" (UID: "bd29f9fc-f6f2-4520-8826-861fa4166a70"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.825098 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd29f9fc-f6f2-4520-8826-861fa4166a70-kube-api-access-56pf8" (OuterVolumeSpecName: "kube-api-access-56pf8") pod "bd29f9fc-f6f2-4520-8826-861fa4166a70" (UID: "bd29f9fc-f6f2-4520-8826-861fa4166a70"). InnerVolumeSpecName "kube-api-access-56pf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:42 crc kubenswrapper[4692]: E0309 09:51:42.832640 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf podName:bd29f9fc-f6f2-4520-8826-861fa4166a70 nodeName:}" failed. No retries permitted until 2026-03-09 09:51:43.332610315 +0000 UTC m=+1904.157345896 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "swiftconf" (UniqueName: "kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf") pod "bd29f9fc-f6f2-4520-8826-861fa4166a70" (UID: "bd29f9fc-f6f2-4520-8826-861fa4166a70") : error deleting /var/lib/kubelet/pods/bd29f9fc-f6f2-4520-8826-861fa4166a70/volume-subpaths: remove /var/lib/kubelet/pods/bd29f9fc-f6f2-4520-8826-861fa4166a70/volume-subpaths: no such file or directory Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.833235 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-scripts" (OuterVolumeSpecName: "scripts") pod "bd29f9fc-f6f2-4520-8826-861fa4166a70" (UID: "bd29f9fc-f6f2-4520-8826-861fa4166a70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.833696 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bd29f9fc-f6f2-4520-8826-861fa4166a70" (UID: "bd29f9fc-f6f2-4520-8826-861fa4166a70"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.912964 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.913243 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56pf8\" (UniqueName: \"kubernetes.io/projected/bd29f9fc-f6f2-4520-8826-861fa4166a70-kube-api-access-56pf8\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.913342 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd29f9fc-f6f2-4520-8826-861fa4166a70-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.913405 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.913459 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd29f9fc-f6f2-4520-8826-861fa4166a70-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.970695 4692 generic.go:334] "Generic (PLEG): container finished" podID="bd29f9fc-f6f2-4520-8826-861fa4166a70" containerID="2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c" exitCode=1 Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.970744 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" event={"ID":"bd29f9fc-f6f2-4520-8826-861fa4166a70","Type":"ContainerDied","Data":"2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c"} Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.970794 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" event={"ID":"bd29f9fc-f6f2-4520-8826-861fa4166a70","Type":"ContainerDied","Data":"4d5d37820e3434f572f1f76434ca3a90e3138be731b570755a7493e1d1b1e9d2"} Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.970820 4692 scope.go:117] "RemoveContainer" containerID="2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.971188 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.990644 4692 scope.go:117] "RemoveContainer" containerID="2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c" Mar 09 09:51:42 crc kubenswrapper[4692]: E0309 09:51:42.991151 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c\": container with ID starting with 2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c not found: ID does not exist" containerID="2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c" Mar 09 09:51:42 crc kubenswrapper[4692]: I0309 09:51:42.991213 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c"} err="failed to get container status \"2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c\": rpc error: code = NotFound desc = could not find container \"2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c\": container with ID starting with 2ba82734342eb59120bb9a835bd96004bdaabf6a2b97d2162ede7a06404fa97c not found: ID does not exist" Mar 09 09:51:43 crc kubenswrapper[4692]: I0309 09:51:43.088943 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx"] Mar 09 09:51:43 crc kubenswrapper[4692]: I0309 09:51:43.094548 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8xrqx"] Mar 09 09:51:43 crc kubenswrapper[4692]: I0309 09:51:43.419947 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf\") pod \"bd29f9fc-f6f2-4520-8826-861fa4166a70\" (UID: \"bd29f9fc-f6f2-4520-8826-861fa4166a70\") " Mar 09 09:51:43 crc kubenswrapper[4692]: I0309 09:51:43.436414 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bd29f9fc-f6f2-4520-8826-861fa4166a70" (UID: "bd29f9fc-f6f2-4520-8826-861fa4166a70"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:51:43 crc kubenswrapper[4692]: I0309 09:51:43.522383 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd29f9fc-f6f2-4520-8826-861fa4166a70-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.079854 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd29f9fc-f6f2-4520-8826-861fa4166a70" path="/var/lib/kubelet/pods/bd29f9fc-f6f2-4520-8826-861fa4166a70/volumes" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267401 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6"] Mar 09 09:51:44 crc kubenswrapper[4692]: E0309 09:51:44.267713 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-server" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267727 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-server" Mar 09 09:51:44 crc kubenswrapper[4692]: E0309 09:51:44.267755 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a69287f-b725-45f6-85ed-9ce460d1e031" containerName="swift-ring-rebalance" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267762 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a69287f-b725-45f6-85ed-9ce460d1e031" containerName="swift-ring-rebalance" Mar 09 09:51:44 crc kubenswrapper[4692]: E0309 09:51:44.267773 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-httpd" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267779 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-httpd" Mar 09 09:51:44 crc kubenswrapper[4692]: E0309 09:51:44.267795 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="extract-content" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267802 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="extract-content" Mar 09 09:51:44 crc kubenswrapper[4692]: E0309 09:51:44.267811 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="extract-utilities" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267818 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="extract-utilities" Mar 09 09:51:44 crc kubenswrapper[4692]: E0309 09:51:44.267830 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd29f9fc-f6f2-4520-8826-861fa4166a70" containerName="swift-ring-rebalance" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267836 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd29f9fc-f6f2-4520-8826-861fa4166a70" containerName="swift-ring-rebalance" Mar 09 09:51:44 crc kubenswrapper[4692]: E0309 09:51:44.267848 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="registry-server" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267854 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="registry-server" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267982 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-httpd" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.267994 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd23df0a-a16c-4a17-8308-909043e28377" containerName="proxy-server" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.268006 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd29f9fc-f6f2-4520-8826-861fa4166a70" containerName="swift-ring-rebalance" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.268019 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="86ec167e-b651-4c2b-87a4-d853a9340409" containerName="registry-server" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.268027 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a69287f-b725-45f6-85ed-9ce460d1e031" containerName="swift-ring-rebalance" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.268500 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.271100 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.271146 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.276606 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6"] Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.435283 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-ring-data-devices\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.435356 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-swiftconf\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.435389 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-scripts\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.435425 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fc7t\" (UniqueName: \"kubernetes.io/projected/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-kube-api-access-7fc7t\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.435476 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-dispersionconf\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.435694 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-etc-swift\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.536980 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-swiftconf\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.537057 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-scripts\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.537107 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fc7t\" (UniqueName: \"kubernetes.io/projected/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-kube-api-access-7fc7t\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.537152 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-dispersionconf\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.537205 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-etc-swift\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.537247 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-ring-data-devices\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.538095 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-ring-data-devices\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.538129 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-etc-swift\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.538455 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-scripts\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.541117 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-swiftconf\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.546973 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-dispersionconf\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.552528 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fc7t\" (UniqueName: \"kubernetes.io/projected/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-kube-api-access-7fc7t\") pod \"swift-ring-rebalance-debug-sqgz6\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:44 crc kubenswrapper[4692]: I0309 09:51:44.585789 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:51:45 crc kubenswrapper[4692]: I0309 09:51:45.013603 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6"] Mar 09 09:51:45 crc kubenswrapper[4692]: I0309 09:51:45.995398 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" event={"ID":"01fe2a63-8b20-4dd1-92de-c8bf25e055d1","Type":"ContainerStarted","Data":"5349ace416b8364a10dc567381bce111843641e77a1df80846e4b20a2a766ddd"} Mar 09 09:51:45 crc kubenswrapper[4692]: I0309 09:51:45.995796 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" event={"ID":"01fe2a63-8b20-4dd1-92de-c8bf25e055d1","Type":"ContainerStarted","Data":"058465c55428889ba34be0d9710314473df825dee487dc626a8b6d0833de84f5"} Mar 09 09:51:46 crc kubenswrapper[4692]: I0309 09:51:46.018656 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" podStartSLOduration=2.018622145 podStartE2EDuration="2.018622145s" podCreationTimestamp="2026-03-09 09:51:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:46.014377654 +0000 UTC m=+1906.839113235" watchObservedRunningTime="2026-03-09 09:51:46.018622145 +0000 UTC m=+1906.843357726" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.073597 4692 generic.go:334] "Generic (PLEG): container finished" podID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerID="ab12c78b3d286b45efb3c2c57c2505919e7441ae58bbc75580420db1252bd058" exitCode=137 Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.073672 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"ab12c78b3d286b45efb3c2c57c2505919e7441ae58bbc75580420db1252bd058"} Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.254653 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.376486 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") pod \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.376563 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lzbn\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-kube-api-access-4lzbn\") pod \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.376626 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-lock\") pod \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.376650 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-cache\") pod \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.376687 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\" (UID: \"c01b6df5-bc9c-4abd-9ce5-e928b06580c7\") " Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.377184 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-lock" (OuterVolumeSpecName: "lock") pod "c01b6df5-bc9c-4abd-9ce5-e928b06580c7" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.377306 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-cache" (OuterVolumeSpecName: "cache") pod "c01b6df5-bc9c-4abd-9ce5-e928b06580c7" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.384553 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-kube-api-access-4lzbn" (OuterVolumeSpecName: "kube-api-access-4lzbn") pod "c01b6df5-bc9c-4abd-9ce5-e928b06580c7" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7"). InnerVolumeSpecName "kube-api-access-4lzbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.384622 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c01b6df5-bc9c-4abd-9ce5-e928b06580c7" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.384612 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "swift") pod "c01b6df5-bc9c-4abd-9ce5-e928b06580c7" (UID: "c01b6df5-bc9c-4abd-9ce5-e928b06580c7"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.478793 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lzbn\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-kube-api-access-4lzbn\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.478839 4692 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-lock\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.478853 4692 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-cache\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.478886 4692 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.478899 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c01b6df5-bc9c-4abd-9ce5-e928b06580c7-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.493298 4692 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 09 09:51:53 crc kubenswrapper[4692]: I0309 09:51:53.580609 4692 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.089194 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"c01b6df5-bc9c-4abd-9ce5-e928b06580c7","Type":"ContainerDied","Data":"77a6f89143fe945e1692feadac74c09e67eb12d755f566307c8fb1e988a0c098"} Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.089526 4692 scope.go:117] "RemoveContainer" containerID="58b3d88d99997185e1fe0c65177b1557bbf0f1600dfcd4c1a78a97dfe4cc3869" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.089740 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.117823 4692 scope.go:117] "RemoveContainer" containerID="ab12c78b3d286b45efb3c2c57c2505919e7441ae58bbc75580420db1252bd058" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.123960 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.129840 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.139621 4692 scope.go:117] "RemoveContainer" containerID="2394ec11824f008a37542215a330958d91bdf6a7d65a1fd73ce74901131426d5" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150509 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150797 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150807 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150826 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150831 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150843 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150849 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150859 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-expirer" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150865 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-expirer" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150874 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="swift-recon-cron" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150879 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="swift-recon-cron" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150893 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150898 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150906 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-server" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150912 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-server" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150925 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150931 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150942 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="rsync" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150948 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="rsync" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150958 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-updater" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150963 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-updater" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150973 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-server" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150978 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-server" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.150987 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-reaper" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.150993 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-reaper" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.151002 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-server" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151007 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-server" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.151017 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-updater" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151023 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-updater" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.151033 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151039 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: E0309 09:51:54.151049 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-sharder" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151055 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-sharder" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151215 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-server" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151228 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-server" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151237 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151244 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151253 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151258 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-reaper" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151269 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-server" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151276 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-updater" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151285 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-updater" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151294 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151302 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="container-sharder" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151311 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-replicator" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151321 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="rsync" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151329 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="object-expirer" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151336 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="swift-recon-cron" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.151343 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" containerName="account-auditor" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.155033 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.175950 4692 scope.go:117] "RemoveContainer" containerID="703b6e2cbdc86106d9e4b0e1d0f94ab5534c785c45615232e8488b6c7384526e" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.179077 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.204175 4692 scope.go:117] "RemoveContainer" containerID="564d437ed65b90cb81f73d2a130ee67b0cedeedb41b6b4d9a24f4f61cfa50bcd" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.224274 4692 scope.go:117] "RemoveContainer" containerID="06be2bd359a325cfb2ef0967a7db004ec3b7c9adb30c68c9aa995111081731ee" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.249926 4692 scope.go:117] "RemoveContainer" containerID="d96fdc6db8ea6a8d4119dbbc1739cb93789eee6ec1375ebdcdbbcaac8fce23fb" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.265856 4692 scope.go:117] "RemoveContainer" containerID="e93f97a75cd47e8689825ca7c8a46036015ce8c783fbd946e4b5aae4c8f79b03" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.281061 4692 scope.go:117] "RemoveContainer" containerID="9c329076d4b5930b205901b8cdceb024a0470b51a44c26609e18b30b6e9ff6b9" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.291349 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-etc-swift\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.291418 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-cache\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.291523 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fswhz\" (UniqueName: \"kubernetes.io/projected/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-kube-api-access-fswhz\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.291663 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-lock\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.291841 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.297787 4692 scope.go:117] "RemoveContainer" containerID="4b04d6a9a6046a7cf5adeb6b3544acc38445e80b0d6ec6bd670e88aac9bee508" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.325873 4692 scope.go:117] "RemoveContainer" containerID="2ea0a775295e7a2ebd710a249b8eef9e502086a8adca32469d0a7e5f0360946b" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.340812 4692 scope.go:117] "RemoveContainer" containerID="3c66beccc4342856159e6f72d1d1ae309389f4f7b64191912f7f8ea930ad529f" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.356412 4692 scope.go:117] "RemoveContainer" containerID="c280851b4e3b0f461e368eb3548350b7ce5623cbea4928739b6a1c9c9b6e4b82" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.371237 4692 scope.go:117] "RemoveContainer" containerID="bb413cb17c31ff233ceedc8a35e86e5b01a697b0c889b4a43599446ef0c3e78d" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.389563 4692 scope.go:117] "RemoveContainer" containerID="c4ecffd2dfe8323c86c3932e7f158bba159aef48ce3a7407fa6ec794dcb21fb9" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.393630 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-cache\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.393678 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fswhz\" (UniqueName: \"kubernetes.io/projected/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-kube-api-access-fswhz\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.393720 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-lock\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.393784 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.393815 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-etc-swift\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.394231 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-cache\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.394544 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-lock\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.394584 4692 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") device mount path \"/mnt/openstack/pv03\"" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.402924 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-etc-swift\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.404794 4692 scope.go:117] "RemoveContainer" containerID="ab65f21b397d18fb36d0d4050eebb38914e73ef3e17f18703e7a399455036458" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.410603 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fswhz\" (UniqueName: \"kubernetes.io/projected/5b3f7c6f-e373-484e-bbe1-1e7b14478b9f-kube-api-access-fswhz\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.413629 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f\") " pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.471665 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-storage-0" Mar 09 09:51:54 crc kubenswrapper[4692]: I0309 09:51:54.929456 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-storage-0"] Mar 09 09:51:55 crc kubenswrapper[4692]: I0309 09:51:55.097713 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"cbfa2fce0f477836f8220fc89c2de850538b18de9dff44125fa95fc532ef3641"} Mar 09 09:51:55 crc kubenswrapper[4692]: I0309 09:51:55.097750 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"4783c8ea0268666ad06a40e8c35cf746bb3f85a2a26d52dd78e25227f061ecf5"} Mar 09 09:51:56 crc kubenswrapper[4692]: I0309 09:51:56.080238 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c01b6df5-bc9c-4abd-9ce5-e928b06580c7" path="/var/lib/kubelet/pods/c01b6df5-bc9c-4abd-9ce5-e928b06580c7/volumes" Mar 09 09:51:56 crc kubenswrapper[4692]: I0309 09:51:56.117587 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"1e14704787dffeb43c88b7e572969e0c72dca54c1455f9fdffdbcc68686e7813"} Mar 09 09:51:56 crc kubenswrapper[4692]: I0309 09:51:56.117628 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"8ae87a52f28314e3e0afcba347f5f0e545f15e76c7d3081dfa0183c1e67716f2"} Mar 09 09:51:56 crc kubenswrapper[4692]: I0309 09:51:56.117638 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"c555239d84adb8911d0eb3ad5cd56a39479a4a21be8b44561695a8e87f08ad4a"} Mar 09 09:51:56 crc kubenswrapper[4692]: I0309 09:51:56.117646 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"ffbb6158991a314fde3c886e5155c93eb1974118da09411c1987a1d4ca11a456"} Mar 09 09:51:56 crc kubenswrapper[4692]: I0309 09:51:56.117654 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"a9b1fb9615da773f1608a68f424cf304ad458699a1a39c362e8b249062389c1f"} Mar 09 09:51:56 crc kubenswrapper[4692]: I0309 09:51:56.117666 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"a960566158e9d9b02a466669778bd6e9899a17707b08ff19e3d57c83bfa5f189"} Mar 09 09:51:57 crc kubenswrapper[4692]: I0309 09:51:57.150879 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"72fe0e95ca3b2ea862953eaf8b4c9e9110e0f3d7ea80535b342c0b37841ad575"} Mar 09 09:51:57 crc kubenswrapper[4692]: I0309 09:51:57.151462 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"4ece2031ca9f9545168ce2f33097995e01af09a908cc631ed1a718c04c80fcdd"} Mar 09 09:51:57 crc kubenswrapper[4692]: I0309 09:51:57.151477 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"16586a62c48f7a1a6bbecb8234b29467c1c77ba37f3e61687c64358907426386"} Mar 09 09:51:57 crc kubenswrapper[4692]: I0309 09:51:57.151488 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"4d9e2212c1271a823949ce1245659cd39061c7058f1cc9f7ab036ca4e23650cd"} Mar 09 09:51:57 crc kubenswrapper[4692]: I0309 09:51:57.151499 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"90a194956a809cbf2df0dc5c24a5cabef7fbff3df653187d1b253f3d5d5b49f2"} Mar 09 09:51:57 crc kubenswrapper[4692]: I0309 09:51:57.151508 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"a3d050c590d3eb9972ad9b1529ccd9fe9cadd5e6cf8a1a6ff9fd09c3013146cd"} Mar 09 09:51:57 crc kubenswrapper[4692]: I0309 09:51:57.151518 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"0d552a053df58d7af92e957e77a9713cd7c4b2c88669aa83ede19d25b6c2a3fd"} Mar 09 09:51:58 crc kubenswrapper[4692]: I0309 09:51:58.165760 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-storage-0" event={"ID":"5b3f7c6f-e373-484e-bbe1-1e7b14478b9f","Type":"ContainerStarted","Data":"7aa5af39ddf9347bcb4790eb84ba90a583122352474dad41948671daa86e0387"} Mar 09 09:51:58 crc kubenswrapper[4692]: I0309 09:51:58.202579 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-storage-0" podStartSLOduration=4.202562859 podStartE2EDuration="4.202562859s" podCreationTimestamp="2026-03-09 09:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:51:58.199446561 +0000 UTC m=+1919.024182162" watchObservedRunningTime="2026-03-09 09:51:58.202562859 +0000 UTC m=+1919.027298450" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.135482 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550832-486cx"] Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.138028 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550832-486cx" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.140193 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.140226 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.141549 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.143463 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550832-486cx"] Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.276900 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmpql\" (UniqueName: \"kubernetes.io/projected/d10c9253-78bc-437f-8dd6-34830d6a263c-kube-api-access-bmpql\") pod \"auto-csr-approver-29550832-486cx\" (UID: \"d10c9253-78bc-437f-8dd6-34830d6a263c\") " pod="openshift-infra/auto-csr-approver-29550832-486cx" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.379032 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmpql\" (UniqueName: \"kubernetes.io/projected/d10c9253-78bc-437f-8dd6-34830d6a263c-kube-api-access-bmpql\") pod \"auto-csr-approver-29550832-486cx\" (UID: \"d10c9253-78bc-437f-8dd6-34830d6a263c\") " pod="openshift-infra/auto-csr-approver-29550832-486cx" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.400066 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmpql\" (UniqueName: \"kubernetes.io/projected/d10c9253-78bc-437f-8dd6-34830d6a263c-kube-api-access-bmpql\") pod \"auto-csr-approver-29550832-486cx\" (UID: \"d10c9253-78bc-437f-8dd6-34830d6a263c\") " pod="openshift-infra/auto-csr-approver-29550832-486cx" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.469942 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550832-486cx" Mar 09 09:52:00 crc kubenswrapper[4692]: I0309 09:52:00.870412 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550832-486cx"] Mar 09 09:52:00 crc kubenswrapper[4692]: W0309 09:52:00.872802 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd10c9253_78bc_437f_8dd6_34830d6a263c.slice/crio-7808dc161aba64a0f3ce8e3549a2ff6ae6a22f4038e1e0345f080535632739b1 WatchSource:0}: Error finding container 7808dc161aba64a0f3ce8e3549a2ff6ae6a22f4038e1e0345f080535632739b1: Status 404 returned error can't find the container with id 7808dc161aba64a0f3ce8e3549a2ff6ae6a22f4038e1e0345f080535632739b1 Mar 09 09:52:01 crc kubenswrapper[4692]: I0309 09:52:01.187970 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550832-486cx" event={"ID":"d10c9253-78bc-437f-8dd6-34830d6a263c","Type":"ContainerStarted","Data":"7808dc161aba64a0f3ce8e3549a2ff6ae6a22f4038e1e0345f080535632739b1"} Mar 09 09:52:03 crc kubenswrapper[4692]: I0309 09:52:03.203030 4692 generic.go:334] "Generic (PLEG): container finished" podID="d10c9253-78bc-437f-8dd6-34830d6a263c" containerID="13a00efb8b3e99458e89f1a1e1ecbaca90137b55c6f8660cb667647a57cf6f2d" exitCode=0 Mar 09 09:52:03 crc kubenswrapper[4692]: I0309 09:52:03.203079 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550832-486cx" event={"ID":"d10c9253-78bc-437f-8dd6-34830d6a263c","Type":"ContainerDied","Data":"13a00efb8b3e99458e89f1a1e1ecbaca90137b55c6f8660cb667647a57cf6f2d"} Mar 09 09:52:03 crc kubenswrapper[4692]: I0309 09:52:03.893714 4692 scope.go:117] "RemoveContainer" containerID="c08c64838317ffe2b9a3715a7ef725eeeed77e577e7eae85c70ab5e550be268b" Mar 09 09:52:04 crc kubenswrapper[4692]: I0309 09:52:04.210908 4692 generic.go:334] "Generic (PLEG): container finished" podID="01fe2a63-8b20-4dd1-92de-c8bf25e055d1" containerID="5349ace416b8364a10dc567381bce111843641e77a1df80846e4b20a2a766ddd" exitCode=0 Mar 09 09:52:04 crc kubenswrapper[4692]: I0309 09:52:04.210989 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" event={"ID":"01fe2a63-8b20-4dd1-92de-c8bf25e055d1","Type":"ContainerDied","Data":"5349ace416b8364a10dc567381bce111843641e77a1df80846e4b20a2a766ddd"} Mar 09 09:52:04 crc kubenswrapper[4692]: I0309 09:52:04.467812 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550832-486cx" Mar 09 09:52:04 crc kubenswrapper[4692]: I0309 09:52:04.536446 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmpql\" (UniqueName: \"kubernetes.io/projected/d10c9253-78bc-437f-8dd6-34830d6a263c-kube-api-access-bmpql\") pod \"d10c9253-78bc-437f-8dd6-34830d6a263c\" (UID: \"d10c9253-78bc-437f-8dd6-34830d6a263c\") " Mar 09 09:52:04 crc kubenswrapper[4692]: I0309 09:52:04.543220 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d10c9253-78bc-437f-8dd6-34830d6a263c-kube-api-access-bmpql" (OuterVolumeSpecName: "kube-api-access-bmpql") pod "d10c9253-78bc-437f-8dd6-34830d6a263c" (UID: "d10c9253-78bc-437f-8dd6-34830d6a263c"). InnerVolumeSpecName "kube-api-access-bmpql". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:04 crc kubenswrapper[4692]: I0309 09:52:04.638732 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmpql\" (UniqueName: \"kubernetes.io/projected/d10c9253-78bc-437f-8dd6-34830d6a263c-kube-api-access-bmpql\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.220650 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550832-486cx" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.220649 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550832-486cx" event={"ID":"d10c9253-78bc-437f-8dd6-34830d6a263c","Type":"ContainerDied","Data":"7808dc161aba64a0f3ce8e3549a2ff6ae6a22f4038e1e0345f080535632739b1"} Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.220849 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7808dc161aba64a0f3ce8e3549a2ff6ae6a22f4038e1e0345f080535632739b1" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.494504 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.554792 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6"] Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.564486 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6"] Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.575006 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550826-9xb8q"] Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.583645 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550826-9xb8q"] Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.653421 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-swiftconf\") pod \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.653820 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-scripts\") pod \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.653957 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-dispersionconf\") pod \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.654141 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-etc-swift\") pod \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.654319 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fc7t\" (UniqueName: \"kubernetes.io/projected/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-kube-api-access-7fc7t\") pod \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.654456 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-ring-data-devices\") pod \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\" (UID: \"01fe2a63-8b20-4dd1-92de-c8bf25e055d1\") " Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.654933 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "01fe2a63-8b20-4dd1-92de-c8bf25e055d1" (UID: "01fe2a63-8b20-4dd1-92de-c8bf25e055d1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.655509 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "01fe2a63-8b20-4dd1-92de-c8bf25e055d1" (UID: "01fe2a63-8b20-4dd1-92de-c8bf25e055d1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.660498 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-kube-api-access-7fc7t" (OuterVolumeSpecName: "kube-api-access-7fc7t") pod "01fe2a63-8b20-4dd1-92de-c8bf25e055d1" (UID: "01fe2a63-8b20-4dd1-92de-c8bf25e055d1"). InnerVolumeSpecName "kube-api-access-7fc7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.675262 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-scripts" (OuterVolumeSpecName: "scripts") pod "01fe2a63-8b20-4dd1-92de-c8bf25e055d1" (UID: "01fe2a63-8b20-4dd1-92de-c8bf25e055d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.681009 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "01fe2a63-8b20-4dd1-92de-c8bf25e055d1" (UID: "01fe2a63-8b20-4dd1-92de-c8bf25e055d1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.683585 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "01fe2a63-8b20-4dd1-92de-c8bf25e055d1" (UID: "01fe2a63-8b20-4dd1-92de-c8bf25e055d1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.756482 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.756522 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.756533 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.756544 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.756552 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.756562 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fc7t\" (UniqueName: \"kubernetes.io/projected/01fe2a63-8b20-4dd1-92de-c8bf25e055d1-kube-api-access-7fc7t\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.947941 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ck62v"] Mar 09 09:52:05 crc kubenswrapper[4692]: E0309 09:52:05.948246 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fe2a63-8b20-4dd1-92de-c8bf25e055d1" containerName="swift-ring-rebalance" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.948257 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fe2a63-8b20-4dd1-92de-c8bf25e055d1" containerName="swift-ring-rebalance" Mar 09 09:52:05 crc kubenswrapper[4692]: E0309 09:52:05.948273 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10c9253-78bc-437f-8dd6-34830d6a263c" containerName="oc" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.948279 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10c9253-78bc-437f-8dd6-34830d6a263c" containerName="oc" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.948429 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d10c9253-78bc-437f-8dd6-34830d6a263c" containerName="oc" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.948447 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="01fe2a63-8b20-4dd1-92de-c8bf25e055d1" containerName="swift-ring-rebalance" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.948910 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:05 crc kubenswrapper[4692]: I0309 09:52:05.967944 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ck62v"] Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.061944 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-ring-data-devices\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.062460 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-dispersionconf\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.062588 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-swiftconf\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.062692 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-scripts\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.062824 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnjf7\" (UniqueName: \"kubernetes.io/projected/bb39311a-050d-4fb6-81d0-6c7423466937-kube-api-access-lnjf7\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.062993 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb39311a-050d-4fb6-81d0-6c7423466937-etc-swift\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.082826 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01fe2a63-8b20-4dd1-92de-c8bf25e055d1" path="/var/lib/kubelet/pods/01fe2a63-8b20-4dd1-92de-c8bf25e055d1/volumes" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.083533 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4684029-a5c7-40e6-9d76-db09d57818e8" path="/var/lib/kubelet/pods/d4684029-a5c7-40e6-9d76-db09d57818e8/volumes" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.165425 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb39311a-050d-4fb6-81d0-6c7423466937-etc-swift\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.165662 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-ring-data-devices\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.165714 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-dispersionconf\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.165746 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-swiftconf\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.165780 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-scripts\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.165853 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnjf7\" (UniqueName: \"kubernetes.io/projected/bb39311a-050d-4fb6-81d0-6c7423466937-kube-api-access-lnjf7\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.165980 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb39311a-050d-4fb6-81d0-6c7423466937-etc-swift\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.167221 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-ring-data-devices\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.168193 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-scripts\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.170489 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-swiftconf\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.170483 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-dispersionconf\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.186000 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnjf7\" (UniqueName: \"kubernetes.io/projected/bb39311a-050d-4fb6-81d0-6c7423466937-kube-api-access-lnjf7\") pod \"swift-ring-rebalance-debug-ck62v\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.230244 4692 scope.go:117] "RemoveContainer" containerID="5349ace416b8364a10dc567381bce111843641e77a1df80846e4b20a2a766ddd" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.230272 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sqgz6" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.267051 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:06 crc kubenswrapper[4692]: I0309 09:52:06.688845 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ck62v"] Mar 09 09:52:07 crc kubenswrapper[4692]: I0309 09:52:07.255998 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" event={"ID":"bb39311a-050d-4fb6-81d0-6c7423466937","Type":"ContainerStarted","Data":"827ef63492e4ffca1170f363aa613f3bbcd16ea3f15663f8cce374696f7a3dce"} Mar 09 09:52:07 crc kubenswrapper[4692]: I0309 09:52:07.256040 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" event={"ID":"bb39311a-050d-4fb6-81d0-6c7423466937","Type":"ContainerStarted","Data":"3e90709b579b17a141286a04cf4c9f8af8a33a6db0ace6d1190e2fead7939696"} Mar 09 09:52:07 crc kubenswrapper[4692]: I0309 09:52:07.276719 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" podStartSLOduration=2.276701752 podStartE2EDuration="2.276701752s" podCreationTimestamp="2026-03-09 09:52:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:07.271060602 +0000 UTC m=+1928.095796193" watchObservedRunningTime="2026-03-09 09:52:07.276701752 +0000 UTC m=+1928.101437333" Mar 09 09:52:08 crc kubenswrapper[4692]: I0309 09:52:08.268135 4692 generic.go:334] "Generic (PLEG): container finished" podID="bb39311a-050d-4fb6-81d0-6c7423466937" containerID="827ef63492e4ffca1170f363aa613f3bbcd16ea3f15663f8cce374696f7a3dce" exitCode=0 Mar 09 09:52:08 crc kubenswrapper[4692]: I0309 09:52:08.268249 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" event={"ID":"bb39311a-050d-4fb6-81d0-6c7423466937","Type":"ContainerDied","Data":"827ef63492e4ffca1170f363aa613f3bbcd16ea3f15663f8cce374696f7a3dce"} Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.541598 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.573933 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ck62v"] Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.579623 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ck62v"] Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.615145 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-dispersionconf\") pod \"bb39311a-050d-4fb6-81d0-6c7423466937\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.615237 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-scripts\") pod \"bb39311a-050d-4fb6-81d0-6c7423466937\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.615260 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-swiftconf\") pod \"bb39311a-050d-4fb6-81d0-6c7423466937\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.615293 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-ring-data-devices\") pod \"bb39311a-050d-4fb6-81d0-6c7423466937\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.615370 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnjf7\" (UniqueName: \"kubernetes.io/projected/bb39311a-050d-4fb6-81d0-6c7423466937-kube-api-access-lnjf7\") pod \"bb39311a-050d-4fb6-81d0-6c7423466937\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.615410 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb39311a-050d-4fb6-81d0-6c7423466937-etc-swift\") pod \"bb39311a-050d-4fb6-81d0-6c7423466937\" (UID: \"bb39311a-050d-4fb6-81d0-6c7423466937\") " Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.616735 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb39311a-050d-4fb6-81d0-6c7423466937-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bb39311a-050d-4fb6-81d0-6c7423466937" (UID: "bb39311a-050d-4fb6-81d0-6c7423466937"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.617494 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bb39311a-050d-4fb6-81d0-6c7423466937" (UID: "bb39311a-050d-4fb6-81d0-6c7423466937"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.625405 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb39311a-050d-4fb6-81d0-6c7423466937-kube-api-access-lnjf7" (OuterVolumeSpecName: "kube-api-access-lnjf7") pod "bb39311a-050d-4fb6-81d0-6c7423466937" (UID: "bb39311a-050d-4fb6-81d0-6c7423466937"). InnerVolumeSpecName "kube-api-access-lnjf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.633518 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-scripts" (OuterVolumeSpecName: "scripts") pod "bb39311a-050d-4fb6-81d0-6c7423466937" (UID: "bb39311a-050d-4fb6-81d0-6c7423466937"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.637449 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bb39311a-050d-4fb6-81d0-6c7423466937" (UID: "bb39311a-050d-4fb6-81d0-6c7423466937"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.637918 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bb39311a-050d-4fb6-81d0-6c7423466937" (UID: "bb39311a-050d-4fb6-81d0-6c7423466937"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.717592 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb39311a-050d-4fb6-81d0-6c7423466937-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.717629 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.717640 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb39311a-050d-4fb6-81d0-6c7423466937-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.717648 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.717657 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb39311a-050d-4fb6-81d0-6c7423466937-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:09 crc kubenswrapper[4692]: I0309 09:52:09.717666 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnjf7\" (UniqueName: \"kubernetes.io/projected/bb39311a-050d-4fb6-81d0-6c7423466937-kube-api-access-lnjf7\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.081254 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb39311a-050d-4fb6-81d0-6c7423466937" path="/var/lib/kubelet/pods/bb39311a-050d-4fb6-81d0-6c7423466937/volumes" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.283222 4692 scope.go:117] "RemoveContainer" containerID="827ef63492e4ffca1170f363aa613f3bbcd16ea3f15663f8cce374696f7a3dce" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.283272 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ck62v" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.720251 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz"] Mar 09 09:52:10 crc kubenswrapper[4692]: E0309 09:52:10.720873 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb39311a-050d-4fb6-81d0-6c7423466937" containerName="swift-ring-rebalance" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.720893 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb39311a-050d-4fb6-81d0-6c7423466937" containerName="swift-ring-rebalance" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.721075 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb39311a-050d-4fb6-81d0-6c7423466937" containerName="swift-ring-rebalance" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.721619 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.725564 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.725764 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.738477 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz"] Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.833043 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-dispersionconf\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.833141 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cc9b193d-6eb5-4f54-b320-414da44f4424-etc-swift\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.833192 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9j84\" (UniqueName: \"kubernetes.io/projected/cc9b193d-6eb5-4f54-b320-414da44f4424-kube-api-access-z9j84\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.833212 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-scripts\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.833231 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-ring-data-devices\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.833254 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-swiftconf\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.934387 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cc9b193d-6eb5-4f54-b320-414da44f4424-etc-swift\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.934449 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9j84\" (UniqueName: \"kubernetes.io/projected/cc9b193d-6eb5-4f54-b320-414da44f4424-kube-api-access-z9j84\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.934478 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-scripts\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.934505 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-ring-data-devices\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.934537 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-swiftconf\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.934573 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-dispersionconf\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.934860 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cc9b193d-6eb5-4f54-b320-414da44f4424-etc-swift\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.935526 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-scripts\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.936154 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-ring-data-devices\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.939765 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-swiftconf\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.939821 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-dispersionconf\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:10 crc kubenswrapper[4692]: I0309 09:52:10.952546 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9j84\" (UniqueName: \"kubernetes.io/projected/cc9b193d-6eb5-4f54-b320-414da44f4424-kube-api-access-z9j84\") pod \"swift-ring-rebalance-debug-lz6kz\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:11 crc kubenswrapper[4692]: I0309 09:52:11.052791 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:11 crc kubenswrapper[4692]: I0309 09:52:11.490442 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz"] Mar 09 09:52:12 crc kubenswrapper[4692]: I0309 09:52:12.299676 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" event={"ID":"cc9b193d-6eb5-4f54-b320-414da44f4424","Type":"ContainerStarted","Data":"209c9fe69490d2c566df58b2617fe4e6708f47d94bf73b96abf6de28064b84dc"} Mar 09 09:52:12 crc kubenswrapper[4692]: I0309 09:52:12.299979 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" event={"ID":"cc9b193d-6eb5-4f54-b320-414da44f4424","Type":"ContainerStarted","Data":"6b1b3b9b6a6e6819a6ba4008e2fe5ab7a215861db1c6254436de2ec92ef10719"} Mar 09 09:52:12 crc kubenswrapper[4692]: I0309 09:52:12.320435 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" podStartSLOduration=2.320412458 podStartE2EDuration="2.320412458s" podCreationTimestamp="2026-03-09 09:52:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:12.317782184 +0000 UTC m=+1933.142517775" watchObservedRunningTime="2026-03-09 09:52:12.320412458 +0000 UTC m=+1933.145148049" Mar 09 09:52:13 crc kubenswrapper[4692]: I0309 09:52:13.308379 4692 generic.go:334] "Generic (PLEG): container finished" podID="cc9b193d-6eb5-4f54-b320-414da44f4424" containerID="209c9fe69490d2c566df58b2617fe4e6708f47d94bf73b96abf6de28064b84dc" exitCode=0 Mar 09 09:52:13 crc kubenswrapper[4692]: I0309 09:52:13.308434 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" event={"ID":"cc9b193d-6eb5-4f54-b320-414da44f4424","Type":"ContainerDied","Data":"209c9fe69490d2c566df58b2617fe4e6708f47d94bf73b96abf6de28064b84dc"} Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.579288 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.610683 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz"] Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.615809 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz"] Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.687332 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9j84\" (UniqueName: \"kubernetes.io/projected/cc9b193d-6eb5-4f54-b320-414da44f4424-kube-api-access-z9j84\") pod \"cc9b193d-6eb5-4f54-b320-414da44f4424\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.687410 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cc9b193d-6eb5-4f54-b320-414da44f4424-etc-swift\") pod \"cc9b193d-6eb5-4f54-b320-414da44f4424\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.687463 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-swiftconf\") pod \"cc9b193d-6eb5-4f54-b320-414da44f4424\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.687536 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-scripts\") pod \"cc9b193d-6eb5-4f54-b320-414da44f4424\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.687577 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-ring-data-devices\") pod \"cc9b193d-6eb5-4f54-b320-414da44f4424\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.687634 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-dispersionconf\") pod \"cc9b193d-6eb5-4f54-b320-414da44f4424\" (UID: \"cc9b193d-6eb5-4f54-b320-414da44f4424\") " Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.688266 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc9b193d-6eb5-4f54-b320-414da44f4424-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cc9b193d-6eb5-4f54-b320-414da44f4424" (UID: "cc9b193d-6eb5-4f54-b320-414da44f4424"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.688642 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cc9b193d-6eb5-4f54-b320-414da44f4424" (UID: "cc9b193d-6eb5-4f54-b320-414da44f4424"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.692639 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc9b193d-6eb5-4f54-b320-414da44f4424-kube-api-access-z9j84" (OuterVolumeSpecName: "kube-api-access-z9j84") pod "cc9b193d-6eb5-4f54-b320-414da44f4424" (UID: "cc9b193d-6eb5-4f54-b320-414da44f4424"). InnerVolumeSpecName "kube-api-access-z9j84". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.709176 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-scripts" (OuterVolumeSpecName: "scripts") pod "cc9b193d-6eb5-4f54-b320-414da44f4424" (UID: "cc9b193d-6eb5-4f54-b320-414da44f4424"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.710128 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cc9b193d-6eb5-4f54-b320-414da44f4424" (UID: "cc9b193d-6eb5-4f54-b320-414da44f4424"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.713220 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cc9b193d-6eb5-4f54-b320-414da44f4424" (UID: "cc9b193d-6eb5-4f54-b320-414da44f4424"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.789691 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.789721 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.789731 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cc9b193d-6eb5-4f54-b320-414da44f4424-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.789743 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cc9b193d-6eb5-4f54-b320-414da44f4424-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.789751 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9j84\" (UniqueName: \"kubernetes.io/projected/cc9b193d-6eb5-4f54-b320-414da44f4424-kube-api-access-z9j84\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:14 crc kubenswrapper[4692]: I0309 09:52:14.789759 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cc9b193d-6eb5-4f54-b320-414da44f4424-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.323600 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b1b3b9b6a6e6819a6ba4008e2fe5ab7a215861db1c6254436de2ec92ef10719" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.323961 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lz6kz" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.777778 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w"] Mar 09 09:52:15 crc kubenswrapper[4692]: E0309 09:52:15.778066 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc9b193d-6eb5-4f54-b320-414da44f4424" containerName="swift-ring-rebalance" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.778078 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc9b193d-6eb5-4f54-b320-414da44f4424" containerName="swift-ring-rebalance" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.778249 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc9b193d-6eb5-4f54-b320-414da44f4424" containerName="swift-ring-rebalance" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.778676 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.780955 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.781529 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.787010 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w"] Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.906954 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.907032 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-scripts\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.907098 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-etc-swift\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.907135 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-swiftconf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.907178 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzkmf\" (UniqueName: \"kubernetes.io/projected/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-kube-api-access-bzkmf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:15 crc kubenswrapper[4692]: I0309 09:52:15.907203 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-ring-data-devices\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.008353 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzkmf\" (UniqueName: \"kubernetes.io/projected/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-kube-api-access-bzkmf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.008395 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-ring-data-devices\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.008465 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.008496 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-scripts\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.008527 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-etc-swift\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.008546 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-swiftconf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.009200 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-etc-swift\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.009420 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-scripts\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.009457 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-ring-data-devices\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.013451 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-swiftconf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.014009 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.027636 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzkmf\" (UniqueName: \"kubernetes.io/projected/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-kube-api-access-bzkmf\") pod \"swift-ring-rebalance-debug-p2q6w\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.082073 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc9b193d-6eb5-4f54-b320-414da44f4424" path="/var/lib/kubelet/pods/cc9b193d-6eb5-4f54-b320-414da44f4424/volumes" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.161556 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:16 crc kubenswrapper[4692]: I0309 09:52:16.602348 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w"] Mar 09 09:52:16 crc kubenswrapper[4692]: W0309 09:52:16.604497 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddea03c17_1ae8_44fe_8b28_23c7bcc3e017.slice/crio-a90a82017c00827a79b0ab875e843c6479ffc17647cfd38786c01c1076e07cb6 WatchSource:0}: Error finding container a90a82017c00827a79b0ab875e843c6479ffc17647cfd38786c01c1076e07cb6: Status 404 returned error can't find the container with id a90a82017c00827a79b0ab875e843c6479ffc17647cfd38786c01c1076e07cb6 Mar 09 09:52:17 crc kubenswrapper[4692]: I0309 09:52:17.341949 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" event={"ID":"dea03c17-1ae8-44fe-8b28-23c7bcc3e017","Type":"ContainerStarted","Data":"7163effcbf5d3400c4ece1b0f4be7f7039b5363fa5dde3aab2c46676ff6d9cae"} Mar 09 09:52:17 crc kubenswrapper[4692]: I0309 09:52:17.342287 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" event={"ID":"dea03c17-1ae8-44fe-8b28-23c7bcc3e017","Type":"ContainerStarted","Data":"a90a82017c00827a79b0ab875e843c6479ffc17647cfd38786c01c1076e07cb6"} Mar 09 09:52:17 crc kubenswrapper[4692]: I0309 09:52:17.372555 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" podStartSLOduration=2.372480571 podStartE2EDuration="2.372480571s" podCreationTimestamp="2026-03-09 09:52:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:17.366771859 +0000 UTC m=+1938.191507450" watchObservedRunningTime="2026-03-09 09:52:17.372480571 +0000 UTC m=+1938.197216152" Mar 09 09:52:18 crc kubenswrapper[4692]: I0309 09:52:18.350397 4692 generic.go:334] "Generic (PLEG): container finished" podID="dea03c17-1ae8-44fe-8b28-23c7bcc3e017" containerID="7163effcbf5d3400c4ece1b0f4be7f7039b5363fa5dde3aab2c46676ff6d9cae" exitCode=0 Mar 09 09:52:18 crc kubenswrapper[4692]: I0309 09:52:18.350489 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" event={"ID":"dea03c17-1ae8-44fe-8b28-23c7bcc3e017","Type":"ContainerDied","Data":"7163effcbf5d3400c4ece1b0f4be7f7039b5363fa5dde3aab2c46676ff6d9cae"} Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.664257 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.699308 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w"] Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.703488 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w"] Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.765500 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf\") pod \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.765881 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzkmf\" (UniqueName: \"kubernetes.io/projected/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-kube-api-access-bzkmf\") pod \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.765920 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-ring-data-devices\") pod \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.765948 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-scripts\") pod \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.765976 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-swiftconf\") pod \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.766009 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-etc-swift\") pod \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.766714 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "dea03c17-1ae8-44fe-8b28-23c7bcc3e017" (UID: "dea03c17-1ae8-44fe-8b28-23c7bcc3e017"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.766938 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "dea03c17-1ae8-44fe-8b28-23c7bcc3e017" (UID: "dea03c17-1ae8-44fe-8b28-23c7bcc3e017"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.776479 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-kube-api-access-bzkmf" (OuterVolumeSpecName: "kube-api-access-bzkmf") pod "dea03c17-1ae8-44fe-8b28-23c7bcc3e017" (UID: "dea03c17-1ae8-44fe-8b28-23c7bcc3e017"). InnerVolumeSpecName "kube-api-access-bzkmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:19 crc kubenswrapper[4692]: E0309 09:52:19.785410 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf podName:dea03c17-1ae8-44fe-8b28-23c7bcc3e017 nodeName:}" failed. No retries permitted until 2026-03-09 09:52:20.285378912 +0000 UTC m=+1941.110114503 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dispersionconf" (UniqueName: "kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf") pod "dea03c17-1ae8-44fe-8b28-23c7bcc3e017" (UID: "dea03c17-1ae8-44fe-8b28-23c7bcc3e017") : error deleting /var/lib/kubelet/pods/dea03c17-1ae8-44fe-8b28-23c7bcc3e017/volume-subpaths: remove /var/lib/kubelet/pods/dea03c17-1ae8-44fe-8b28-23c7bcc3e017/volume-subpaths: no such file or directory Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.785920 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-scripts" (OuterVolumeSpecName: "scripts") pod "dea03c17-1ae8-44fe-8b28-23c7bcc3e017" (UID: "dea03c17-1ae8-44fe-8b28-23c7bcc3e017"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.789433 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "dea03c17-1ae8-44fe-8b28-23c7bcc3e017" (UID: "dea03c17-1ae8-44fe-8b28-23c7bcc3e017"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.867389 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzkmf\" (UniqueName: \"kubernetes.io/projected/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-kube-api-access-bzkmf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.867428 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.867438 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.867449 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:19 crc kubenswrapper[4692]: I0309 09:52:19.867458 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.364640 4692 scope.go:117] "RemoveContainer" containerID="7163effcbf5d3400c4ece1b0f4be7f7039b5363fa5dde3aab2c46676ff6d9cae" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.364659 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p2q6w" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.373822 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf\") pod \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\" (UID: \"dea03c17-1ae8-44fe-8b28-23c7bcc3e017\") " Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.377248 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "dea03c17-1ae8-44fe-8b28-23c7bcc3e017" (UID: "dea03c17-1ae8-44fe-8b28-23c7bcc3e017"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.475581 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/dea03c17-1ae8-44fe-8b28-23c7bcc3e017-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.847547 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xv44"] Mar 09 09:52:20 crc kubenswrapper[4692]: E0309 09:52:20.848744 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea03c17-1ae8-44fe-8b28-23c7bcc3e017" containerName="swift-ring-rebalance" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.848840 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea03c17-1ae8-44fe-8b28-23c7bcc3e017" containerName="swift-ring-rebalance" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.849085 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea03c17-1ae8-44fe-8b28-23c7bcc3e017" containerName="swift-ring-rebalance" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.849721 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.851719 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.851759 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.853708 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xv44"] Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.982259 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-scripts\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.982474 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3f6cffd-1e52-4c01-9e88-141bee954435-etc-swift\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.982585 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjgmw\" (UniqueName: \"kubernetes.io/projected/d3f6cffd-1e52-4c01-9e88-141bee954435-kube-api-access-tjgmw\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.982739 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-dispersionconf\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.982842 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-swiftconf\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:20 crc kubenswrapper[4692]: I0309 09:52:20.982925 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-ring-data-devices\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.084088 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3f6cffd-1e52-4c01-9e88-141bee954435-etc-swift\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.084149 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgmw\" (UniqueName: \"kubernetes.io/projected/d3f6cffd-1e52-4c01-9e88-141bee954435-kube-api-access-tjgmw\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.084202 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-dispersionconf\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.084229 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-swiftconf\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.084253 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-ring-data-devices\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.084276 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-scripts\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.084614 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3f6cffd-1e52-4c01-9e88-141bee954435-etc-swift\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.085061 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-scripts\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.085192 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-ring-data-devices\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.087918 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-dispersionconf\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.088603 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-swiftconf\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.111410 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgmw\" (UniqueName: \"kubernetes.io/projected/d3f6cffd-1e52-4c01-9e88-141bee954435-kube-api-access-tjgmw\") pod \"swift-ring-rebalance-debug-7xv44\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.170034 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:21 crc kubenswrapper[4692]: I0309 09:52:21.615481 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xv44"] Mar 09 09:52:21 crc kubenswrapper[4692]: W0309 09:52:21.617969 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3f6cffd_1e52_4c01_9e88_141bee954435.slice/crio-6b11cb4653112e69ad022a0c2b7a6e3f11f455827e3e69e481b6be0f86825faf WatchSource:0}: Error finding container 6b11cb4653112e69ad022a0c2b7a6e3f11f455827e3e69e481b6be0f86825faf: Status 404 returned error can't find the container with id 6b11cb4653112e69ad022a0c2b7a6e3f11f455827e3e69e481b6be0f86825faf Mar 09 09:52:22 crc kubenswrapper[4692]: I0309 09:52:22.080295 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea03c17-1ae8-44fe-8b28-23c7bcc3e017" path="/var/lib/kubelet/pods/dea03c17-1ae8-44fe-8b28-23c7bcc3e017/volumes" Mar 09 09:52:22 crc kubenswrapper[4692]: I0309 09:52:22.383003 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" event={"ID":"d3f6cffd-1e52-4c01-9e88-141bee954435","Type":"ContainerStarted","Data":"4517cad5bcb4f554212542d663de526a9492034948b19012e9a59f0451a37bf9"} Mar 09 09:52:22 crc kubenswrapper[4692]: I0309 09:52:22.383340 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" event={"ID":"d3f6cffd-1e52-4c01-9e88-141bee954435","Type":"ContainerStarted","Data":"6b11cb4653112e69ad022a0c2b7a6e3f11f455827e3e69e481b6be0f86825faf"} Mar 09 09:52:22 crc kubenswrapper[4692]: I0309 09:52:22.409402 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" podStartSLOduration=2.409385938 podStartE2EDuration="2.409385938s" podCreationTimestamp="2026-03-09 09:52:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:22.407876495 +0000 UTC m=+1943.232612096" watchObservedRunningTime="2026-03-09 09:52:22.409385938 +0000 UTC m=+1943.234121519" Mar 09 09:52:23 crc kubenswrapper[4692]: I0309 09:52:23.391055 4692 generic.go:334] "Generic (PLEG): container finished" podID="d3f6cffd-1e52-4c01-9e88-141bee954435" containerID="4517cad5bcb4f554212542d663de526a9492034948b19012e9a59f0451a37bf9" exitCode=0 Mar 09 09:52:23 crc kubenswrapper[4692]: I0309 09:52:23.391100 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" event={"ID":"d3f6cffd-1e52-4c01-9e88-141bee954435","Type":"ContainerDied","Data":"4517cad5bcb4f554212542d663de526a9492034948b19012e9a59f0451a37bf9"} Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.686254 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.727384 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xv44"] Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.732439 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7xv44"] Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.846038 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-swiftconf\") pod \"d3f6cffd-1e52-4c01-9e88-141bee954435\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.846133 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-scripts\") pod \"d3f6cffd-1e52-4c01-9e88-141bee954435\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.846213 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjgmw\" (UniqueName: \"kubernetes.io/projected/d3f6cffd-1e52-4c01-9e88-141bee954435-kube-api-access-tjgmw\") pod \"d3f6cffd-1e52-4c01-9e88-141bee954435\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.846237 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3f6cffd-1e52-4c01-9e88-141bee954435-etc-swift\") pod \"d3f6cffd-1e52-4c01-9e88-141bee954435\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.846258 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-ring-data-devices\") pod \"d3f6cffd-1e52-4c01-9e88-141bee954435\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.846303 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-dispersionconf\") pod \"d3f6cffd-1e52-4c01-9e88-141bee954435\" (UID: \"d3f6cffd-1e52-4c01-9e88-141bee954435\") " Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.846891 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d3f6cffd-1e52-4c01-9e88-141bee954435" (UID: "d3f6cffd-1e52-4c01-9e88-141bee954435"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.847130 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3f6cffd-1e52-4c01-9e88-141bee954435-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d3f6cffd-1e52-4c01-9e88-141bee954435" (UID: "d3f6cffd-1e52-4c01-9e88-141bee954435"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.851751 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f6cffd-1e52-4c01-9e88-141bee954435-kube-api-access-tjgmw" (OuterVolumeSpecName: "kube-api-access-tjgmw") pod "d3f6cffd-1e52-4c01-9e88-141bee954435" (UID: "d3f6cffd-1e52-4c01-9e88-141bee954435"). InnerVolumeSpecName "kube-api-access-tjgmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.864785 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-scripts" (OuterVolumeSpecName: "scripts") pod "d3f6cffd-1e52-4c01-9e88-141bee954435" (UID: "d3f6cffd-1e52-4c01-9e88-141bee954435"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.867417 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d3f6cffd-1e52-4c01-9e88-141bee954435" (UID: "d3f6cffd-1e52-4c01-9e88-141bee954435"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.870013 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d3f6cffd-1e52-4c01-9e88-141bee954435" (UID: "d3f6cffd-1e52-4c01-9e88-141bee954435"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.948416 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.948654 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.948664 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjgmw\" (UniqueName: \"kubernetes.io/projected/d3f6cffd-1e52-4c01-9e88-141bee954435-kube-api-access-tjgmw\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.948677 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d3f6cffd-1e52-4c01-9e88-141bee954435-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.948686 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d3f6cffd-1e52-4c01-9e88-141bee954435-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:24 crc kubenswrapper[4692]: I0309 09:52:24.948694 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d3f6cffd-1e52-4c01-9e88-141bee954435-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.408727 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b11cb4653112e69ad022a0c2b7a6e3f11f455827e3e69e481b6be0f86825faf" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.408771 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7xv44" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.891379 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7nlml"] Mar 09 09:52:25 crc kubenswrapper[4692]: E0309 09:52:25.891786 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f6cffd-1e52-4c01-9e88-141bee954435" containerName="swift-ring-rebalance" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.891802 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f6cffd-1e52-4c01-9e88-141bee954435" containerName="swift-ring-rebalance" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.891994 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f6cffd-1e52-4c01-9e88-141bee954435" containerName="swift-ring-rebalance" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.892665 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.894438 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.894492 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.899130 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7nlml"] Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.962418 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-swiftconf\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.962483 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdr5h\" (UniqueName: \"kubernetes.io/projected/6321c92b-9e0b-461c-91ee-2987b367ff35-kube-api-access-fdr5h\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.962529 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-dispersionconf\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.962696 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6321c92b-9e0b-461c-91ee-2987b367ff35-etc-swift\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.962750 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-ring-data-devices\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:25 crc kubenswrapper[4692]: I0309 09:52:25.962854 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-scripts\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.064022 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-dispersionconf\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.064081 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6321c92b-9e0b-461c-91ee-2987b367ff35-etc-swift\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.064098 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-ring-data-devices\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.064145 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-scripts\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.064197 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-swiftconf\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.064239 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdr5h\" (UniqueName: \"kubernetes.io/projected/6321c92b-9e0b-461c-91ee-2987b367ff35-kube-api-access-fdr5h\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.064905 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6321c92b-9e0b-461c-91ee-2987b367ff35-etc-swift\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.065226 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-ring-data-devices\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.065307 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-scripts\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.068593 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-dispersionconf\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.068974 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-swiftconf\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.081194 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdr5h\" (UniqueName: \"kubernetes.io/projected/6321c92b-9e0b-461c-91ee-2987b367ff35-kube-api-access-fdr5h\") pod \"swift-ring-rebalance-debug-7nlml\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.081477 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f6cffd-1e52-4c01-9e88-141bee954435" path="/var/lib/kubelet/pods/d3f6cffd-1e52-4c01-9e88-141bee954435/volumes" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.210500 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:26 crc kubenswrapper[4692]: I0309 09:52:26.606075 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7nlml"] Mar 09 09:52:26 crc kubenswrapper[4692]: W0309 09:52:26.607382 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6321c92b_9e0b_461c_91ee_2987b367ff35.slice/crio-9c5f601228adece496eca6c18d2b83e35519f246d55ab6ad4289c845fcd3186d WatchSource:0}: Error finding container 9c5f601228adece496eca6c18d2b83e35519f246d55ab6ad4289c845fcd3186d: Status 404 returned error can't find the container with id 9c5f601228adece496eca6c18d2b83e35519f246d55ab6ad4289c845fcd3186d Mar 09 09:52:27 crc kubenswrapper[4692]: I0309 09:52:27.424672 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" event={"ID":"6321c92b-9e0b-461c-91ee-2987b367ff35","Type":"ContainerStarted","Data":"687076ab81d25ed26aaaedf23186cf0ecd744914ada9a44358671564c86039ca"} Mar 09 09:52:27 crc kubenswrapper[4692]: I0309 09:52:27.424939 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" event={"ID":"6321c92b-9e0b-461c-91ee-2987b367ff35","Type":"ContainerStarted","Data":"9c5f601228adece496eca6c18d2b83e35519f246d55ab6ad4289c845fcd3186d"} Mar 09 09:52:28 crc kubenswrapper[4692]: I0309 09:52:28.432527 4692 generic.go:334] "Generic (PLEG): container finished" podID="6321c92b-9e0b-461c-91ee-2987b367ff35" containerID="687076ab81d25ed26aaaedf23186cf0ecd744914ada9a44358671564c86039ca" exitCode=0 Mar 09 09:52:28 crc kubenswrapper[4692]: I0309 09:52:28.432648 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" event={"ID":"6321c92b-9e0b-461c-91ee-2987b367ff35","Type":"ContainerDied","Data":"687076ab81d25ed26aaaedf23186cf0ecd744914ada9a44358671564c86039ca"} Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.743556 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.787690 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7nlml"] Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.793978 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7nlml"] Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.814597 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdr5h\" (UniqueName: \"kubernetes.io/projected/6321c92b-9e0b-461c-91ee-2987b367ff35-kube-api-access-fdr5h\") pod \"6321c92b-9e0b-461c-91ee-2987b367ff35\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.814698 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-ring-data-devices\") pod \"6321c92b-9e0b-461c-91ee-2987b367ff35\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.814734 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-dispersionconf\") pod \"6321c92b-9e0b-461c-91ee-2987b367ff35\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.814796 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6321c92b-9e0b-461c-91ee-2987b367ff35-etc-swift\") pod \"6321c92b-9e0b-461c-91ee-2987b367ff35\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.814832 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-swiftconf\") pod \"6321c92b-9e0b-461c-91ee-2987b367ff35\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.814878 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-scripts\") pod \"6321c92b-9e0b-461c-91ee-2987b367ff35\" (UID: \"6321c92b-9e0b-461c-91ee-2987b367ff35\") " Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.815103 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6321c92b-9e0b-461c-91ee-2987b367ff35" (UID: "6321c92b-9e0b-461c-91ee-2987b367ff35"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.815242 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.815863 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6321c92b-9e0b-461c-91ee-2987b367ff35-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6321c92b-9e0b-461c-91ee-2987b367ff35" (UID: "6321c92b-9e0b-461c-91ee-2987b367ff35"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.820275 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6321c92b-9e0b-461c-91ee-2987b367ff35-kube-api-access-fdr5h" (OuterVolumeSpecName: "kube-api-access-fdr5h") pod "6321c92b-9e0b-461c-91ee-2987b367ff35" (UID: "6321c92b-9e0b-461c-91ee-2987b367ff35"). InnerVolumeSpecName "kube-api-access-fdr5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.836242 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-scripts" (OuterVolumeSpecName: "scripts") pod "6321c92b-9e0b-461c-91ee-2987b367ff35" (UID: "6321c92b-9e0b-461c-91ee-2987b367ff35"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.837098 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6321c92b-9e0b-461c-91ee-2987b367ff35" (UID: "6321c92b-9e0b-461c-91ee-2987b367ff35"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.841019 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6321c92b-9e0b-461c-91ee-2987b367ff35" (UID: "6321c92b-9e0b-461c-91ee-2987b367ff35"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.916544 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.916602 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6321c92b-9e0b-461c-91ee-2987b367ff35-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.916619 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdr5h\" (UniqueName: \"kubernetes.io/projected/6321c92b-9e0b-461c-91ee-2987b367ff35-kube-api-access-fdr5h\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.916635 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6321c92b-9e0b-461c-91ee-2987b367ff35-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:29 crc kubenswrapper[4692]: I0309 09:52:29.916646 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6321c92b-9e0b-461c-91ee-2987b367ff35-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.081316 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6321c92b-9e0b-461c-91ee-2987b367ff35" path="/var/lib/kubelet/pods/6321c92b-9e0b-461c-91ee-2987b367ff35/volumes" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.448766 4692 scope.go:117] "RemoveContainer" containerID="687076ab81d25ed26aaaedf23186cf0ecd744914ada9a44358671564c86039ca" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.448819 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7nlml" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.935940 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7"] Mar 09 09:52:30 crc kubenswrapper[4692]: E0309 09:52:30.936266 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6321c92b-9e0b-461c-91ee-2987b367ff35" containerName="swift-ring-rebalance" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.936279 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6321c92b-9e0b-461c-91ee-2987b367ff35" containerName="swift-ring-rebalance" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.936448 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6321c92b-9e0b-461c-91ee-2987b367ff35" containerName="swift-ring-rebalance" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.937020 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.939345 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.939657 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:30 crc kubenswrapper[4692]: I0309 09:52:30.946000 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7"] Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.030870 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8e81d69-48cf-479a-a97b-dd0dad84c1de-etc-swift\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.031236 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-ring-data-devices\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.031273 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-scripts\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.031292 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-dispersionconf\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.031325 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-swiftconf\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.031369 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4drv\" (UniqueName: \"kubernetes.io/projected/d8e81d69-48cf-479a-a97b-dd0dad84c1de-kube-api-access-m4drv\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.132524 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-scripts\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.132571 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-dispersionconf\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.132613 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-swiftconf\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.132652 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4drv\" (UniqueName: \"kubernetes.io/projected/d8e81d69-48cf-479a-a97b-dd0dad84c1de-kube-api-access-m4drv\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.132729 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8e81d69-48cf-479a-a97b-dd0dad84c1de-etc-swift\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.132756 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-ring-data-devices\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.133614 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-ring-data-devices\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.133968 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8e81d69-48cf-479a-a97b-dd0dad84c1de-etc-swift\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.135563 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-scripts\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.144665 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-dispersionconf\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.147462 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-swiftconf\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.150646 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4drv\" (UniqueName: \"kubernetes.io/projected/d8e81d69-48cf-479a-a97b-dd0dad84c1de-kube-api-access-m4drv\") pod \"swift-ring-rebalance-debug-7ncl7\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.263741 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:31 crc kubenswrapper[4692]: I0309 09:52:31.662620 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7"] Mar 09 09:52:31 crc kubenswrapper[4692]: W0309 09:52:31.671202 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8e81d69_48cf_479a_a97b_dd0dad84c1de.slice/crio-a0a84edd2ed4cc83feb67b24c190ffc5e5f2b862e75e10563797b5acc8f4faef WatchSource:0}: Error finding container a0a84edd2ed4cc83feb67b24c190ffc5e5f2b862e75e10563797b5acc8f4faef: Status 404 returned error can't find the container with id a0a84edd2ed4cc83feb67b24c190ffc5e5f2b862e75e10563797b5acc8f4faef Mar 09 09:52:32 crc kubenswrapper[4692]: I0309 09:52:32.467398 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" event={"ID":"d8e81d69-48cf-479a-a97b-dd0dad84c1de","Type":"ContainerStarted","Data":"96195b47de9285eeebb5eafe3541bf95f85f1b855a554db9c83ee01c5bd7b5ea"} Mar 09 09:52:32 crc kubenswrapper[4692]: I0309 09:52:32.467740 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" event={"ID":"d8e81d69-48cf-479a-a97b-dd0dad84c1de","Type":"ContainerStarted","Data":"a0a84edd2ed4cc83feb67b24c190ffc5e5f2b862e75e10563797b5acc8f4faef"} Mar 09 09:52:32 crc kubenswrapper[4692]: I0309 09:52:32.496626 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" podStartSLOduration=2.4965785670000002 podStartE2EDuration="2.496578567s" podCreationTimestamp="2026-03-09 09:52:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:32.48538966 +0000 UTC m=+1953.310125261" watchObservedRunningTime="2026-03-09 09:52:32.496578567 +0000 UTC m=+1953.321314158" Mar 09 09:52:33 crc kubenswrapper[4692]: I0309 09:52:33.475455 4692 generic.go:334] "Generic (PLEG): container finished" podID="d8e81d69-48cf-479a-a97b-dd0dad84c1de" containerID="96195b47de9285eeebb5eafe3541bf95f85f1b855a554db9c83ee01c5bd7b5ea" exitCode=0 Mar 09 09:52:33 crc kubenswrapper[4692]: I0309 09:52:33.475494 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" event={"ID":"d8e81d69-48cf-479a-a97b-dd0dad84c1de","Type":"ContainerDied","Data":"96195b47de9285eeebb5eafe3541bf95f85f1b855a554db9c83ee01c5bd7b5ea"} Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.763754 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.799617 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7"] Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.805948 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7"] Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.896911 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-ring-data-devices\") pod \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.896979 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-scripts\") pod \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.897069 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-swiftconf\") pod \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.897116 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8e81d69-48cf-479a-a97b-dd0dad84c1de-etc-swift\") pod \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.897177 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-dispersionconf\") pod \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.897205 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4drv\" (UniqueName: \"kubernetes.io/projected/d8e81d69-48cf-479a-a97b-dd0dad84c1de-kube-api-access-m4drv\") pod \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\" (UID: \"d8e81d69-48cf-479a-a97b-dd0dad84c1de\") " Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.897902 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d8e81d69-48cf-479a-a97b-dd0dad84c1de" (UID: "d8e81d69-48cf-479a-a97b-dd0dad84c1de"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.898407 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8e81d69-48cf-479a-a97b-dd0dad84c1de-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d8e81d69-48cf-479a-a97b-dd0dad84c1de" (UID: "d8e81d69-48cf-479a-a97b-dd0dad84c1de"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.905620 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8e81d69-48cf-479a-a97b-dd0dad84c1de-kube-api-access-m4drv" (OuterVolumeSpecName: "kube-api-access-m4drv") pod "d8e81d69-48cf-479a-a97b-dd0dad84c1de" (UID: "d8e81d69-48cf-479a-a97b-dd0dad84c1de"). InnerVolumeSpecName "kube-api-access-m4drv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.919506 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d8e81d69-48cf-479a-a97b-dd0dad84c1de" (UID: "d8e81d69-48cf-479a-a97b-dd0dad84c1de"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.920471 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-scripts" (OuterVolumeSpecName: "scripts") pod "d8e81d69-48cf-479a-a97b-dd0dad84c1de" (UID: "d8e81d69-48cf-479a-a97b-dd0dad84c1de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.920565 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d8e81d69-48cf-479a-a97b-dd0dad84c1de" (UID: "d8e81d69-48cf-479a-a97b-dd0dad84c1de"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.998878 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.999183 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d8e81d69-48cf-479a-a97b-dd0dad84c1de-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.999294 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d8e81d69-48cf-479a-a97b-dd0dad84c1de-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.999355 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4drv\" (UniqueName: \"kubernetes.io/projected/d8e81d69-48cf-479a-a97b-dd0dad84c1de-kube-api-access-m4drv\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.999413 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:34 crc kubenswrapper[4692]: I0309 09:52:34.999469 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d8e81d69-48cf-479a-a97b-dd0dad84c1de-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.492334 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0a84edd2ed4cc83feb67b24c190ffc5e5f2b862e75e10563797b5acc8f4faef" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.492389 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7ncl7" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.948358 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7"] Mar 09 09:52:35 crc kubenswrapper[4692]: E0309 09:52:35.948649 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8e81d69-48cf-479a-a97b-dd0dad84c1de" containerName="swift-ring-rebalance" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.948661 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8e81d69-48cf-479a-a97b-dd0dad84c1de" containerName="swift-ring-rebalance" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.948803 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8e81d69-48cf-479a-a97b-dd0dad84c1de" containerName="swift-ring-rebalance" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.949302 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.951434 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.951651 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:35 crc kubenswrapper[4692]: I0309 09:52:35.963357 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7"] Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.014378 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a59a3401-a8e3-4d4f-a832-217f7d774324-etc-swift\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.014431 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-swiftconf\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.014465 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l8cz\" (UniqueName: \"kubernetes.io/projected/a59a3401-a8e3-4d4f-a832-217f7d774324-kube-api-access-6l8cz\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.014574 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-scripts\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.014786 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-ring-data-devices\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.015085 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-dispersionconf\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.081037 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8e81d69-48cf-479a-a97b-dd0dad84c1de" path="/var/lib/kubelet/pods/d8e81d69-48cf-479a-a97b-dd0dad84c1de/volumes" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.117287 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a59a3401-a8e3-4d4f-a832-217f7d774324-etc-swift\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.117342 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-swiftconf\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.117388 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l8cz\" (UniqueName: \"kubernetes.io/projected/a59a3401-a8e3-4d4f-a832-217f7d774324-kube-api-access-6l8cz\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.117421 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-scripts\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.117445 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-ring-data-devices\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.117533 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-dispersionconf\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.118090 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a59a3401-a8e3-4d4f-a832-217f7d774324-etc-swift\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.119081 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-scripts\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.119132 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-ring-data-devices\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.123827 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-swiftconf\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.125064 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-dispersionconf\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.137624 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l8cz\" (UniqueName: \"kubernetes.io/projected/a59a3401-a8e3-4d4f-a832-217f7d774324-kube-api-access-6l8cz\") pod \"swift-ring-rebalance-debug-kjhc7\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.274639 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:36 crc kubenswrapper[4692]: I0309 09:52:36.694674 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7"] Mar 09 09:52:36 crc kubenswrapper[4692]: W0309 09:52:36.699553 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda59a3401_a8e3_4d4f_a832_217f7d774324.slice/crio-b6e678c41576040571436a4eab713b2f722806f16db15bbf060eec783b147d2b WatchSource:0}: Error finding container b6e678c41576040571436a4eab713b2f722806f16db15bbf060eec783b147d2b: Status 404 returned error can't find the container with id b6e678c41576040571436a4eab713b2f722806f16db15bbf060eec783b147d2b Mar 09 09:52:37 crc kubenswrapper[4692]: I0309 09:52:37.507835 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" event={"ID":"a59a3401-a8e3-4d4f-a832-217f7d774324","Type":"ContainerStarted","Data":"d649c9d9308eba17c6594cb7218f6d4107806b6dccab4ba5963d4d477cb5ce1c"} Mar 09 09:52:37 crc kubenswrapper[4692]: I0309 09:52:37.508133 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" event={"ID":"a59a3401-a8e3-4d4f-a832-217f7d774324","Type":"ContainerStarted","Data":"b6e678c41576040571436a4eab713b2f722806f16db15bbf060eec783b147d2b"} Mar 09 09:52:37 crc kubenswrapper[4692]: I0309 09:52:37.526095 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" podStartSLOduration=2.52607671 podStartE2EDuration="2.52607671s" podCreationTimestamp="2026-03-09 09:52:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:37.523685453 +0000 UTC m=+1958.348421034" watchObservedRunningTime="2026-03-09 09:52:37.52607671 +0000 UTC m=+1958.350812291" Mar 09 09:52:38 crc kubenswrapper[4692]: I0309 09:52:38.516859 4692 generic.go:334] "Generic (PLEG): container finished" podID="a59a3401-a8e3-4d4f-a832-217f7d774324" containerID="d649c9d9308eba17c6594cb7218f6d4107806b6dccab4ba5963d4d477cb5ce1c" exitCode=0 Mar 09 09:52:38 crc kubenswrapper[4692]: I0309 09:52:38.516934 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" event={"ID":"a59a3401-a8e3-4d4f-a832-217f7d774324","Type":"ContainerDied","Data":"d649c9d9308eba17c6594cb7218f6d4107806b6dccab4ba5963d4d477cb5ce1c"} Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.759628 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.797012 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7"] Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.802813 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7"] Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.877066 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-ring-data-devices\") pod \"a59a3401-a8e3-4d4f-a832-217f7d774324\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.877206 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-dispersionconf\") pod \"a59a3401-a8e3-4d4f-a832-217f7d774324\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.877253 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l8cz\" (UniqueName: \"kubernetes.io/projected/a59a3401-a8e3-4d4f-a832-217f7d774324-kube-api-access-6l8cz\") pod \"a59a3401-a8e3-4d4f-a832-217f7d774324\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.877289 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a59a3401-a8e3-4d4f-a832-217f7d774324-etc-swift\") pod \"a59a3401-a8e3-4d4f-a832-217f7d774324\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.877325 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-swiftconf\") pod \"a59a3401-a8e3-4d4f-a832-217f7d774324\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.877356 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-scripts\") pod \"a59a3401-a8e3-4d4f-a832-217f7d774324\" (UID: \"a59a3401-a8e3-4d4f-a832-217f7d774324\") " Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.877767 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a59a3401-a8e3-4d4f-a832-217f7d774324" (UID: "a59a3401-a8e3-4d4f-a832-217f7d774324"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.878022 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a59a3401-a8e3-4d4f-a832-217f7d774324-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a59a3401-a8e3-4d4f-a832-217f7d774324" (UID: "a59a3401-a8e3-4d4f-a832-217f7d774324"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.887363 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a59a3401-a8e3-4d4f-a832-217f7d774324-kube-api-access-6l8cz" (OuterVolumeSpecName: "kube-api-access-6l8cz") pod "a59a3401-a8e3-4d4f-a832-217f7d774324" (UID: "a59a3401-a8e3-4d4f-a832-217f7d774324"). InnerVolumeSpecName "kube-api-access-6l8cz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.900935 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a59a3401-a8e3-4d4f-a832-217f7d774324" (UID: "a59a3401-a8e3-4d4f-a832-217f7d774324"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.902787 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a59a3401-a8e3-4d4f-a832-217f7d774324" (UID: "a59a3401-a8e3-4d4f-a832-217f7d774324"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.903370 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-scripts" (OuterVolumeSpecName: "scripts") pod "a59a3401-a8e3-4d4f-a832-217f7d774324" (UID: "a59a3401-a8e3-4d4f-a832-217f7d774324"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.979850 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.979916 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a59a3401-a8e3-4d4f-a832-217f7d774324-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.979929 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.979940 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l8cz\" (UniqueName: \"kubernetes.io/projected/a59a3401-a8e3-4d4f-a832-217f7d774324-kube-api-access-6l8cz\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.979953 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a59a3401-a8e3-4d4f-a832-217f7d774324-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:39 crc kubenswrapper[4692]: I0309 09:52:39.979962 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a59a3401-a8e3-4d4f-a832-217f7d774324-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.080091 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a59a3401-a8e3-4d4f-a832-217f7d774324" path="/var/lib/kubelet/pods/a59a3401-a8e3-4d4f-a832-217f7d774324/volumes" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.531570 4692 scope.go:117] "RemoveContainer" containerID="d649c9d9308eba17c6594cb7218f6d4107806b6dccab4ba5963d4d477cb5ce1c" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.531682 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kjhc7" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.955935 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-868rl"] Mar 09 09:52:40 crc kubenswrapper[4692]: E0309 09:52:40.956305 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a59a3401-a8e3-4d4f-a832-217f7d774324" containerName="swift-ring-rebalance" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.956321 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a59a3401-a8e3-4d4f-a832-217f7d774324" containerName="swift-ring-rebalance" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.956491 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a59a3401-a8e3-4d4f-a832-217f7d774324" containerName="swift-ring-rebalance" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.957043 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.963525 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.963527 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:40 crc kubenswrapper[4692]: I0309 09:52:40.970368 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-868rl"] Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.095878 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-scripts\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.095940 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-ring-data-devices\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.096067 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-swiftconf\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.096130 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvt4g\" (UniqueName: \"kubernetes.io/projected/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-kube-api-access-fvt4g\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.096220 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-etc-swift\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.096302 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-dispersionconf\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197087 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-scripts\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197187 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-ring-data-devices\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197252 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-swiftconf\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197289 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvt4g\" (UniqueName: \"kubernetes.io/projected/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-kube-api-access-fvt4g\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197319 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-etc-swift\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197350 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-dispersionconf\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197942 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-ring-data-devices\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.197980 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-etc-swift\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.198381 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-scripts\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.202074 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-swiftconf\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.208152 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-dispersionconf\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.215472 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvt4g\" (UniqueName: \"kubernetes.io/projected/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-kube-api-access-fvt4g\") pod \"swift-ring-rebalance-debug-868rl\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.274272 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:41 crc kubenswrapper[4692]: I0309 09:52:41.661729 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-868rl"] Mar 09 09:52:41 crc kubenswrapper[4692]: W0309 09:52:41.668454 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22c2a812_c9cc_40ae_bbe5_f69f6338ea3b.slice/crio-2cce04291f9e2ab3cadb59c0fc66178f1a94bfb8880b793af44f23700d15dcdb WatchSource:0}: Error finding container 2cce04291f9e2ab3cadb59c0fc66178f1a94bfb8880b793af44f23700d15dcdb: Status 404 returned error can't find the container with id 2cce04291f9e2ab3cadb59c0fc66178f1a94bfb8880b793af44f23700d15dcdb Mar 09 09:52:42 crc kubenswrapper[4692]: I0309 09:52:42.549832 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" event={"ID":"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b","Type":"ContainerStarted","Data":"3580d1770362f5765d81516c9e8b50f135dd442db13b96320b675908b45da6f5"} Mar 09 09:52:42 crc kubenswrapper[4692]: I0309 09:52:42.550430 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" event={"ID":"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b","Type":"ContainerStarted","Data":"2cce04291f9e2ab3cadb59c0fc66178f1a94bfb8880b793af44f23700d15dcdb"} Mar 09 09:52:42 crc kubenswrapper[4692]: I0309 09:52:42.570279 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" podStartSLOduration=2.570240549 podStartE2EDuration="2.570240549s" podCreationTimestamp="2026-03-09 09:52:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:42.563127198 +0000 UTC m=+1963.387862779" watchObservedRunningTime="2026-03-09 09:52:42.570240549 +0000 UTC m=+1963.394976130" Mar 09 09:52:43 crc kubenswrapper[4692]: I0309 09:52:43.562287 4692 generic.go:334] "Generic (PLEG): container finished" podID="22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" containerID="3580d1770362f5765d81516c9e8b50f135dd442db13b96320b675908b45da6f5" exitCode=0 Mar 09 09:52:43 crc kubenswrapper[4692]: I0309 09:52:43.562615 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" event={"ID":"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b","Type":"ContainerDied","Data":"3580d1770362f5765d81516c9e8b50f135dd442db13b96320b675908b45da6f5"} Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.819526 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.855828 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-868rl"] Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.860942 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-868rl"] Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.950004 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-ring-data-devices\") pod \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.950141 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvt4g\" (UniqueName: \"kubernetes.io/projected/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-kube-api-access-fvt4g\") pod \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.950243 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-swiftconf\") pod \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.950285 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-scripts\") pod \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.950319 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-dispersionconf\") pod \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.950379 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-etc-swift\") pod \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\" (UID: \"22c2a812-c9cc-40ae-bbe5-f69f6338ea3b\") " Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.950664 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" (UID: "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.951349 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" (UID: "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.955474 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-kube-api-access-fvt4g" (OuterVolumeSpecName: "kube-api-access-fvt4g") pod "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" (UID: "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b"). InnerVolumeSpecName "kube-api-access-fvt4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.968661 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-scripts" (OuterVolumeSpecName: "scripts") pod "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" (UID: "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.970396 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" (UID: "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:44 crc kubenswrapper[4692]: I0309 09:52:44.978842 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" (UID: "22c2a812-c9cc-40ae-bbe5-f69f6338ea3b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.052392 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.052430 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvt4g\" (UniqueName: \"kubernetes.io/projected/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-kube-api-access-fvt4g\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.052447 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.052458 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.052469 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.052480 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.580252 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2cce04291f9e2ab3cadb59c0fc66178f1a94bfb8880b793af44f23700d15dcdb" Mar 09 09:52:45 crc kubenswrapper[4692]: I0309 09:52:45.580295 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-868rl" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.018646 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn"] Mar 09 09:52:46 crc kubenswrapper[4692]: E0309 09:52:46.019134 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" containerName="swift-ring-rebalance" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.019151 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" containerName="swift-ring-rebalance" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.019350 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" containerName="swift-ring-rebalance" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.020019 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.022212 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.022431 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.038804 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn"] Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.065755 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-swiftconf\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.065817 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-dispersionconf\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.065848 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2v2rs\" (UniqueName: \"kubernetes.io/projected/05ac4813-1b1e-4985-9a43-30108f71eeb8-kube-api-access-2v2rs\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.065870 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-ring-data-devices\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.065939 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/05ac4813-1b1e-4985-9a43-30108f71eeb8-etc-swift\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.065977 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-scripts\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.079900 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c2a812-c9cc-40ae-bbe5-f69f6338ea3b" path="/var/lib/kubelet/pods/22c2a812-c9cc-40ae-bbe5-f69f6338ea3b/volumes" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.167328 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-ring-data-devices\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.167397 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-scripts\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.167422 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/05ac4813-1b1e-4985-9a43-30108f71eeb8-etc-swift\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.167559 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-swiftconf\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.167607 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-dispersionconf\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.167654 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2v2rs\" (UniqueName: \"kubernetes.io/projected/05ac4813-1b1e-4985-9a43-30108f71eeb8-kube-api-access-2v2rs\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.168777 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-ring-data-devices\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.168848 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-scripts\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.169137 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/05ac4813-1b1e-4985-9a43-30108f71eeb8-etc-swift\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.172685 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-dispersionconf\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.172697 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-swiftconf\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.184553 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2v2rs\" (UniqueName: \"kubernetes.io/projected/05ac4813-1b1e-4985-9a43-30108f71eeb8-kube-api-access-2v2rs\") pod \"swift-ring-rebalance-debug-jz4wn\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.336719 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.561884 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn"] Mar 09 09:52:46 crc kubenswrapper[4692]: W0309 09:52:46.568559 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05ac4813_1b1e_4985_9a43_30108f71eeb8.slice/crio-5f566bb77061f8c432d4dc5e900f9345f2f26fa9758a3ca1006a00ed466c0d25 WatchSource:0}: Error finding container 5f566bb77061f8c432d4dc5e900f9345f2f26fa9758a3ca1006a00ed466c0d25: Status 404 returned error can't find the container with id 5f566bb77061f8c432d4dc5e900f9345f2f26fa9758a3ca1006a00ed466c0d25 Mar 09 09:52:46 crc kubenswrapper[4692]: I0309 09:52:46.595109 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" event={"ID":"05ac4813-1b1e-4985-9a43-30108f71eeb8","Type":"ContainerStarted","Data":"5f566bb77061f8c432d4dc5e900f9345f2f26fa9758a3ca1006a00ed466c0d25"} Mar 09 09:52:47 crc kubenswrapper[4692]: I0309 09:52:47.603653 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" event={"ID":"05ac4813-1b1e-4985-9a43-30108f71eeb8","Type":"ContainerStarted","Data":"6b609de2f27727acaced98a32b07bf775fdcab92e22252a7f23619909ae78a6b"} Mar 09 09:52:47 crc kubenswrapper[4692]: I0309 09:52:47.641748 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" podStartSLOduration=2.641727942 podStartE2EDuration="2.641727942s" podCreationTimestamp="2026-03-09 09:52:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:47.629923567 +0000 UTC m=+1968.454659158" watchObservedRunningTime="2026-03-09 09:52:47.641727942 +0000 UTC m=+1968.466463523" Mar 09 09:52:48 crc kubenswrapper[4692]: I0309 09:52:48.612527 4692 generic.go:334] "Generic (PLEG): container finished" podID="05ac4813-1b1e-4985-9a43-30108f71eeb8" containerID="6b609de2f27727acaced98a32b07bf775fdcab92e22252a7f23619909ae78a6b" exitCode=0 Mar 09 09:52:48 crc kubenswrapper[4692]: I0309 09:52:48.612610 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" event={"ID":"05ac4813-1b1e-4985-9a43-30108f71eeb8","Type":"ContainerDied","Data":"6b609de2f27727acaced98a32b07bf775fdcab92e22252a7f23619909ae78a6b"} Mar 09 09:52:49 crc kubenswrapper[4692]: I0309 09:52:49.923900 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:49 crc kubenswrapper[4692]: I0309 09:52:49.973689 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn"] Mar 09 09:52:49 crc kubenswrapper[4692]: I0309 09:52:49.980963 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn"] Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.022631 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-swiftconf\") pod \"05ac4813-1b1e-4985-9a43-30108f71eeb8\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.022834 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-scripts\") pod \"05ac4813-1b1e-4985-9a43-30108f71eeb8\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.022868 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/05ac4813-1b1e-4985-9a43-30108f71eeb8-etc-swift\") pod \"05ac4813-1b1e-4985-9a43-30108f71eeb8\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.022906 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2v2rs\" (UniqueName: \"kubernetes.io/projected/05ac4813-1b1e-4985-9a43-30108f71eeb8-kube-api-access-2v2rs\") pod \"05ac4813-1b1e-4985-9a43-30108f71eeb8\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.022997 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-ring-data-devices\") pod \"05ac4813-1b1e-4985-9a43-30108f71eeb8\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.023039 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-dispersionconf\") pod \"05ac4813-1b1e-4985-9a43-30108f71eeb8\" (UID: \"05ac4813-1b1e-4985-9a43-30108f71eeb8\") " Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.024844 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "05ac4813-1b1e-4985-9a43-30108f71eeb8" (UID: "05ac4813-1b1e-4985-9a43-30108f71eeb8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.025273 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05ac4813-1b1e-4985-9a43-30108f71eeb8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "05ac4813-1b1e-4985-9a43-30108f71eeb8" (UID: "05ac4813-1b1e-4985-9a43-30108f71eeb8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.030919 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05ac4813-1b1e-4985-9a43-30108f71eeb8-kube-api-access-2v2rs" (OuterVolumeSpecName: "kube-api-access-2v2rs") pod "05ac4813-1b1e-4985-9a43-30108f71eeb8" (UID: "05ac4813-1b1e-4985-9a43-30108f71eeb8"). InnerVolumeSpecName "kube-api-access-2v2rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.048359 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-scripts" (OuterVolumeSpecName: "scripts") pod "05ac4813-1b1e-4985-9a43-30108f71eeb8" (UID: "05ac4813-1b1e-4985-9a43-30108f71eeb8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.050771 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "05ac4813-1b1e-4985-9a43-30108f71eeb8" (UID: "05ac4813-1b1e-4985-9a43-30108f71eeb8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.051129 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "05ac4813-1b1e-4985-9a43-30108f71eeb8" (UID: "05ac4813-1b1e-4985-9a43-30108f71eeb8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.082353 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05ac4813-1b1e-4985-9a43-30108f71eeb8" path="/var/lib/kubelet/pods/05ac4813-1b1e-4985-9a43-30108f71eeb8/volumes" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.125280 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.125321 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.125338 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/05ac4813-1b1e-4985-9a43-30108f71eeb8-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.125348 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/05ac4813-1b1e-4985-9a43-30108f71eeb8-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.125357 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/05ac4813-1b1e-4985-9a43-30108f71eeb8-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.125365 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2v2rs\" (UniqueName: \"kubernetes.io/projected/05ac4813-1b1e-4985-9a43-30108f71eeb8-kube-api-access-2v2rs\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.631156 4692 scope.go:117] "RemoveContainer" containerID="6b609de2f27727acaced98a32b07bf775fdcab92e22252a7f23619909ae78a6b" Mar 09 09:52:50 crc kubenswrapper[4692]: I0309 09:52:50.631241 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jz4wn" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.132402 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dclgw"] Mar 09 09:52:51 crc kubenswrapper[4692]: E0309 09:52:51.132819 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ac4813-1b1e-4985-9a43-30108f71eeb8" containerName="swift-ring-rebalance" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.132835 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ac4813-1b1e-4985-9a43-30108f71eeb8" containerName="swift-ring-rebalance" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.133113 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ac4813-1b1e-4985-9a43-30108f71eeb8" containerName="swift-ring-rebalance" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.133791 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.136358 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.137830 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.154617 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dclgw"] Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.240219 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-dispersionconf\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.240375 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-scripts\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.240703 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-ring-data-devices\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.240792 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/92155596-fd18-406b-a34c-65e97ad02669-etc-swift\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.240940 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-swiftconf\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.240993 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnh96\" (UniqueName: \"kubernetes.io/projected/92155596-fd18-406b-a34c-65e97ad02669-kube-api-access-cnh96\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.342502 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-ring-data-devices\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.342572 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/92155596-fd18-406b-a34c-65e97ad02669-etc-swift\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.342610 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-swiftconf\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.342634 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnh96\" (UniqueName: \"kubernetes.io/projected/92155596-fd18-406b-a34c-65e97ad02669-kube-api-access-cnh96\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.342737 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-dispersionconf\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.342781 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-scripts\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.343360 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/92155596-fd18-406b-a34c-65e97ad02669-etc-swift\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.344029 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-scripts\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.344067 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-ring-data-devices\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.350226 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-swiftconf\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.353732 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-dispersionconf\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.367713 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnh96\" (UniqueName: \"kubernetes.io/projected/92155596-fd18-406b-a34c-65e97ad02669-kube-api-access-cnh96\") pod \"swift-ring-rebalance-debug-dclgw\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.455410 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:51 crc kubenswrapper[4692]: I0309 09:52:51.894556 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dclgw"] Mar 09 09:52:52 crc kubenswrapper[4692]: I0309 09:52:52.685657 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" event={"ID":"92155596-fd18-406b-a34c-65e97ad02669","Type":"ContainerStarted","Data":"09b01c5676c4a787f0f0da4a2d13ab72cd517be381ded7a3b8734bd995dfe863"} Mar 09 09:52:52 crc kubenswrapper[4692]: I0309 09:52:52.687436 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" event={"ID":"92155596-fd18-406b-a34c-65e97ad02669","Type":"ContainerStarted","Data":"42d93d957575437f225913b5720e0b2d17c33dffc7ac7948f1a332a097d26b3f"} Mar 09 09:52:52 crc kubenswrapper[4692]: I0309 09:52:52.712129 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" podStartSLOduration=1.712108373 podStartE2EDuration="1.712108373s" podCreationTimestamp="2026-03-09 09:52:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:52.71024207 +0000 UTC m=+1973.534977671" watchObservedRunningTime="2026-03-09 09:52:52.712108373 +0000 UTC m=+1973.536843964" Mar 09 09:52:53 crc kubenswrapper[4692]: I0309 09:52:53.697634 4692 generic.go:334] "Generic (PLEG): container finished" podID="92155596-fd18-406b-a34c-65e97ad02669" containerID="09b01c5676c4a787f0f0da4a2d13ab72cd517be381ded7a3b8734bd995dfe863" exitCode=0 Mar 09 09:52:53 crc kubenswrapper[4692]: I0309 09:52:53.697683 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" event={"ID":"92155596-fd18-406b-a34c-65e97ad02669","Type":"ContainerDied","Data":"09b01c5676c4a787f0f0da4a2d13ab72cd517be381ded7a3b8734bd995dfe863"} Mar 09 09:52:54 crc kubenswrapper[4692]: I0309 09:52:54.994716 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.028412 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dclgw"] Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.033891 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dclgw"] Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.103068 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-swiftconf\") pod \"92155596-fd18-406b-a34c-65e97ad02669\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.103148 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-ring-data-devices\") pod \"92155596-fd18-406b-a34c-65e97ad02669\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.103234 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnh96\" (UniqueName: \"kubernetes.io/projected/92155596-fd18-406b-a34c-65e97ad02669-kube-api-access-cnh96\") pod \"92155596-fd18-406b-a34c-65e97ad02669\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.103276 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/92155596-fd18-406b-a34c-65e97ad02669-etc-swift\") pod \"92155596-fd18-406b-a34c-65e97ad02669\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.103329 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-scripts\") pod \"92155596-fd18-406b-a34c-65e97ad02669\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.103406 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-dispersionconf\") pod \"92155596-fd18-406b-a34c-65e97ad02669\" (UID: \"92155596-fd18-406b-a34c-65e97ad02669\") " Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.104113 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "92155596-fd18-406b-a34c-65e97ad02669" (UID: "92155596-fd18-406b-a34c-65e97ad02669"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.104485 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92155596-fd18-406b-a34c-65e97ad02669-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "92155596-fd18-406b-a34c-65e97ad02669" (UID: "92155596-fd18-406b-a34c-65e97ad02669"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.109597 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92155596-fd18-406b-a34c-65e97ad02669-kube-api-access-cnh96" (OuterVolumeSpecName: "kube-api-access-cnh96") pod "92155596-fd18-406b-a34c-65e97ad02669" (UID: "92155596-fd18-406b-a34c-65e97ad02669"). InnerVolumeSpecName "kube-api-access-cnh96". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.125887 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-scripts" (OuterVolumeSpecName: "scripts") pod "92155596-fd18-406b-a34c-65e97ad02669" (UID: "92155596-fd18-406b-a34c-65e97ad02669"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.126253 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "92155596-fd18-406b-a34c-65e97ad02669" (UID: "92155596-fd18-406b-a34c-65e97ad02669"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.130330 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "92155596-fd18-406b-a34c-65e97ad02669" (UID: "92155596-fd18-406b-a34c-65e97ad02669"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.204936 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnh96\" (UniqueName: \"kubernetes.io/projected/92155596-fd18-406b-a34c-65e97ad02669-kube-api-access-cnh96\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.204971 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/92155596-fd18-406b-a34c-65e97ad02669-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.204985 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.204994 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.205002 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/92155596-fd18-406b-a34c-65e97ad02669-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.205011 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/92155596-fd18-406b-a34c-65e97ad02669-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.714022 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42d93d957575437f225913b5720e0b2d17c33dffc7ac7948f1a332a097d26b3f" Mar 09 09:52:55 crc kubenswrapper[4692]: I0309 09:52:55.714092 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dclgw" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.081538 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92155596-fd18-406b-a34c-65e97ad02669" path="/var/lib/kubelet/pods/92155596-fd18-406b-a34c-65e97ad02669/volumes" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.204765 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf"] Mar 09 09:52:56 crc kubenswrapper[4692]: E0309 09:52:56.205077 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92155596-fd18-406b-a34c-65e97ad02669" containerName="swift-ring-rebalance" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.205098 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="92155596-fd18-406b-a34c-65e97ad02669" containerName="swift-ring-rebalance" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.205283 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="92155596-fd18-406b-a34c-65e97ad02669" containerName="swift-ring-rebalance" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.205730 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.212627 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.212931 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.217495 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf"] Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.323881 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bml79\" (UniqueName: \"kubernetes.io/projected/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-kube-api-access-bml79\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.324009 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-scripts\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.324212 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-ring-data-devices\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.324254 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-etc-swift\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.324368 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-dispersionconf\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.324457 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-swiftconf\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.426314 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bml79\" (UniqueName: \"kubernetes.io/projected/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-kube-api-access-bml79\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.426370 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-scripts\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.426399 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-ring-data-devices\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.426424 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-etc-swift\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.426471 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-dispersionconf\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.426517 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-swiftconf\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.427076 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-etc-swift\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.427374 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-scripts\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.427607 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-ring-data-devices\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.430614 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-swiftconf\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.434544 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-dispersionconf\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.444721 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bml79\" (UniqueName: \"kubernetes.io/projected/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-kube-api-access-bml79\") pod \"swift-ring-rebalance-debug-s4nhf\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.527329 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:52:56 crc kubenswrapper[4692]: I0309 09:52:56.744367 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf"] Mar 09 09:52:56 crc kubenswrapper[4692]: W0309 09:52:56.748543 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ed6f445_e9c0_4fe7_85be_fd8ae221a693.slice/crio-9ffe086f4d02fe50456f1e6e2ed567e377350ee12dc2bc51c34664371fe13b7b WatchSource:0}: Error finding container 9ffe086f4d02fe50456f1e6e2ed567e377350ee12dc2bc51c34664371fe13b7b: Status 404 returned error can't find the container with id 9ffe086f4d02fe50456f1e6e2ed567e377350ee12dc2bc51c34664371fe13b7b Mar 09 09:52:57 crc kubenswrapper[4692]: I0309 09:52:57.735718 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" event={"ID":"9ed6f445-e9c0-4fe7-85be-fd8ae221a693","Type":"ContainerStarted","Data":"d799bc2f3ab555eb5cb84a47c334a308cfc48e8020a40fb7504c2b2d4115d171"} Mar 09 09:52:57 crc kubenswrapper[4692]: I0309 09:52:57.736055 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" event={"ID":"9ed6f445-e9c0-4fe7-85be-fd8ae221a693","Type":"ContainerStarted","Data":"9ffe086f4d02fe50456f1e6e2ed567e377350ee12dc2bc51c34664371fe13b7b"} Mar 09 09:52:57 crc kubenswrapper[4692]: I0309 09:52:57.752108 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" podStartSLOduration=1.752089374 podStartE2EDuration="1.752089374s" podCreationTimestamp="2026-03-09 09:52:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:52:57.750566341 +0000 UTC m=+1978.575301922" watchObservedRunningTime="2026-03-09 09:52:57.752089374 +0000 UTC m=+1978.576824965" Mar 09 09:52:58 crc kubenswrapper[4692]: I0309 09:52:58.743146 4692 generic.go:334] "Generic (PLEG): container finished" podID="9ed6f445-e9c0-4fe7-85be-fd8ae221a693" containerID="d799bc2f3ab555eb5cb84a47c334a308cfc48e8020a40fb7504c2b2d4115d171" exitCode=0 Mar 09 09:52:58 crc kubenswrapper[4692]: I0309 09:52:58.743212 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" event={"ID":"9ed6f445-e9c0-4fe7-85be-fd8ae221a693","Type":"ContainerDied","Data":"d799bc2f3ab555eb5cb84a47c334a308cfc48e8020a40fb7504c2b2d4115d171"} Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.022223 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.064415 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf"] Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.068737 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf"] Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.182945 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-etc-swift\") pod \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.183294 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-dispersionconf\") pod \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.183391 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-scripts\") pod \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.183523 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bml79\" (UniqueName: \"kubernetes.io/projected/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-kube-api-access-bml79\") pod \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.183682 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9ed6f445-e9c0-4fe7-85be-fd8ae221a693" (UID: "9ed6f445-e9c0-4fe7-85be-fd8ae221a693"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.183783 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-ring-data-devices\") pod \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.183871 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-swiftconf\") pod \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\" (UID: \"9ed6f445-e9c0-4fe7-85be-fd8ae221a693\") " Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.184220 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.184330 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "9ed6f445-e9c0-4fe7-85be-fd8ae221a693" (UID: "9ed6f445-e9c0-4fe7-85be-fd8ae221a693"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.192380 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-kube-api-access-bml79" (OuterVolumeSpecName: "kube-api-access-bml79") pod "9ed6f445-e9c0-4fe7-85be-fd8ae221a693" (UID: "9ed6f445-e9c0-4fe7-85be-fd8ae221a693"). InnerVolumeSpecName "kube-api-access-bml79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.205738 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-scripts" (OuterVolumeSpecName: "scripts") pod "9ed6f445-e9c0-4fe7-85be-fd8ae221a693" (UID: "9ed6f445-e9c0-4fe7-85be-fd8ae221a693"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.207524 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "9ed6f445-e9c0-4fe7-85be-fd8ae221a693" (UID: "9ed6f445-e9c0-4fe7-85be-fd8ae221a693"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.215730 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "9ed6f445-e9c0-4fe7-85be-fd8ae221a693" (UID: "9ed6f445-e9c0-4fe7-85be-fd8ae221a693"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.285864 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.285899 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.285909 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bml79\" (UniqueName: \"kubernetes.io/projected/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-kube-api-access-bml79\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.285920 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.285929 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/9ed6f445-e9c0-4fe7-85be-fd8ae221a693-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.761232 4692 scope.go:117] "RemoveContainer" containerID="d799bc2f3ab555eb5cb84a47c334a308cfc48e8020a40fb7504c2b2d4115d171" Mar 09 09:53:00 crc kubenswrapper[4692]: I0309 09:53:00.761265 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s4nhf" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.197065 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr"] Mar 09 09:53:01 crc kubenswrapper[4692]: E0309 09:53:01.197729 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed6f445-e9c0-4fe7-85be-fd8ae221a693" containerName="swift-ring-rebalance" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.197744 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed6f445-e9c0-4fe7-85be-fd8ae221a693" containerName="swift-ring-rebalance" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.197915 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed6f445-e9c0-4fe7-85be-fd8ae221a693" containerName="swift-ring-rebalance" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.198501 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.204811 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.204820 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.206927 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr"] Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.302663 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-etc-swift\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.302724 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-swiftconf\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.302754 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-scripts\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.302831 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-ring-data-devices\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.302879 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-dispersionconf\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.303073 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r687b\" (UniqueName: \"kubernetes.io/projected/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-kube-api-access-r687b\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.404219 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r687b\" (UniqueName: \"kubernetes.io/projected/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-kube-api-access-r687b\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.404307 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-etc-swift\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.404337 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-swiftconf\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.404361 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-scripts\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.404409 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-ring-data-devices\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.404428 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-dispersionconf\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.404760 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-etc-swift\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.405312 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-scripts\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.405355 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-ring-data-devices\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.412616 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-swiftconf\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.412626 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-dispersionconf\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.419560 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r687b\" (UniqueName: \"kubernetes.io/projected/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-kube-api-access-r687b\") pod \"swift-ring-rebalance-debug-rmhxr\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.531483 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:01 crc kubenswrapper[4692]: I0309 09:53:01.964777 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr"] Mar 09 09:53:01 crc kubenswrapper[4692]: W0309 09:53:01.967961 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1901b37b_1ffa_4ca8_8819_d597b6f8ab13.slice/crio-88cd23e79dc5b61b9dc2db20337a5434d7a02ba062b2c265073941eec6534c2b WatchSource:0}: Error finding container 88cd23e79dc5b61b9dc2db20337a5434d7a02ba062b2c265073941eec6534c2b: Status 404 returned error can't find the container with id 88cd23e79dc5b61b9dc2db20337a5434d7a02ba062b2c265073941eec6534c2b Mar 09 09:53:02 crc kubenswrapper[4692]: I0309 09:53:02.082361 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ed6f445-e9c0-4fe7-85be-fd8ae221a693" path="/var/lib/kubelet/pods/9ed6f445-e9c0-4fe7-85be-fd8ae221a693/volumes" Mar 09 09:53:02 crc kubenswrapper[4692]: I0309 09:53:02.777729 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" event={"ID":"1901b37b-1ffa-4ca8-8819-d597b6f8ab13","Type":"ContainerStarted","Data":"a7480c36d0baf51cfa88845f2aec3a3e54a7632d4c9e48c893bfbf48500cd0af"} Mar 09 09:53:02 crc kubenswrapper[4692]: I0309 09:53:02.778027 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" event={"ID":"1901b37b-1ffa-4ca8-8819-d597b6f8ab13","Type":"ContainerStarted","Data":"88cd23e79dc5b61b9dc2db20337a5434d7a02ba062b2c265073941eec6534c2b"} Mar 09 09:53:02 crc kubenswrapper[4692]: I0309 09:53:02.795091 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" podStartSLOduration=1.795074279 podStartE2EDuration="1.795074279s" podCreationTimestamp="2026-03-09 09:53:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:02.792726473 +0000 UTC m=+1983.617462074" watchObservedRunningTime="2026-03-09 09:53:02.795074279 +0000 UTC m=+1983.619809860" Mar 09 09:53:03 crc kubenswrapper[4692]: I0309 09:53:03.790619 4692 generic.go:334] "Generic (PLEG): container finished" podID="1901b37b-1ffa-4ca8-8819-d597b6f8ab13" containerID="a7480c36d0baf51cfa88845f2aec3a3e54a7632d4c9e48c893bfbf48500cd0af" exitCode=0 Mar 09 09:53:03 crc kubenswrapper[4692]: I0309 09:53:03.790966 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" event={"ID":"1901b37b-1ffa-4ca8-8819-d597b6f8ab13","Type":"ContainerDied","Data":"a7480c36d0baf51cfa88845f2aec3a3e54a7632d4c9e48c893bfbf48500cd0af"} Mar 09 09:53:03 crc kubenswrapper[4692]: I0309 09:53:03.994062 4692 scope.go:117] "RemoveContainer" containerID="40d197987b30e35706f6500edf302b8d02e84068e88e2caabb4b823448007189" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.069387 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.101703 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr"] Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.107866 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr"] Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.164628 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r687b\" (UniqueName: \"kubernetes.io/projected/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-kube-api-access-r687b\") pod \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.164693 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-scripts\") pod \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.164776 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-swiftconf\") pod \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.164808 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-etc-swift\") pod \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.164836 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-ring-data-devices\") pod \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.164857 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-dispersionconf\") pod \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\" (UID: \"1901b37b-1ffa-4ca8-8819-d597b6f8ab13\") " Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.166273 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1901b37b-1ffa-4ca8-8819-d597b6f8ab13" (UID: "1901b37b-1ffa-4ca8-8819-d597b6f8ab13"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.167382 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1901b37b-1ffa-4ca8-8819-d597b6f8ab13" (UID: "1901b37b-1ffa-4ca8-8819-d597b6f8ab13"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.171407 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-kube-api-access-r687b" (OuterVolumeSpecName: "kube-api-access-r687b") pod "1901b37b-1ffa-4ca8-8819-d597b6f8ab13" (UID: "1901b37b-1ffa-4ca8-8819-d597b6f8ab13"). InnerVolumeSpecName "kube-api-access-r687b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.185699 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-scripts" (OuterVolumeSpecName: "scripts") pod "1901b37b-1ffa-4ca8-8819-d597b6f8ab13" (UID: "1901b37b-1ffa-4ca8-8819-d597b6f8ab13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.187516 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1901b37b-1ffa-4ca8-8819-d597b6f8ab13" (UID: "1901b37b-1ffa-4ca8-8819-d597b6f8ab13"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.187576 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1901b37b-1ffa-4ca8-8819-d597b6f8ab13" (UID: "1901b37b-1ffa-4ca8-8819-d597b6f8ab13"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.267122 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r687b\" (UniqueName: \"kubernetes.io/projected/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-kube-api-access-r687b\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.267159 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.267183 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.267193 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.267203 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.267212 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1901b37b-1ffa-4ca8-8819-d597b6f8ab13-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.808117 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88cd23e79dc5b61b9dc2db20337a5434d7a02ba062b2c265073941eec6534c2b" Mar 09 09:53:05 crc kubenswrapper[4692]: I0309 09:53:05.808152 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rmhxr" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.090659 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1901b37b-1ffa-4ca8-8819-d597b6f8ab13" path="/var/lib/kubelet/pods/1901b37b-1ffa-4ca8-8819-d597b6f8ab13/volumes" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.268085 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pb77z"] Mar 09 09:53:06 crc kubenswrapper[4692]: E0309 09:53:06.268404 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1901b37b-1ffa-4ca8-8819-d597b6f8ab13" containerName="swift-ring-rebalance" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.268420 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1901b37b-1ffa-4ca8-8819-d597b6f8ab13" containerName="swift-ring-rebalance" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.268680 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1901b37b-1ffa-4ca8-8819-d597b6f8ab13" containerName="swift-ring-rebalance" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.269288 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.272006 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.272149 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.283871 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pb77z"] Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.382530 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-ring-data-devices\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.382591 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz4hr\" (UniqueName: \"kubernetes.io/projected/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-kube-api-access-xz4hr\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.382631 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-etc-swift\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.382660 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-swiftconf\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.382685 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-dispersionconf\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.382950 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-scripts\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.484517 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-ring-data-devices\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.484595 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz4hr\" (UniqueName: \"kubernetes.io/projected/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-kube-api-access-xz4hr\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.484626 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-etc-swift\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.484655 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-swiftconf\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.484672 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-dispersionconf\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.484705 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-scripts\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.485230 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-ring-data-devices\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.485471 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-etc-swift\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.486120 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-scripts\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.493666 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-swiftconf\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.494846 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-dispersionconf\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.502355 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz4hr\" (UniqueName: \"kubernetes.io/projected/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-kube-api-access-xz4hr\") pod \"swift-ring-rebalance-debug-pb77z\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:06 crc kubenswrapper[4692]: I0309 09:53:06.594536 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:07 crc kubenswrapper[4692]: I0309 09:53:07.002948 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pb77z"] Mar 09 09:53:07 crc kubenswrapper[4692]: I0309 09:53:07.823050 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" event={"ID":"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4","Type":"ContainerStarted","Data":"9770ea0091fa66c133afeb08906c31ff59c3bdc74f88c85deca0b7b375da682b"} Mar 09 09:53:07 crc kubenswrapper[4692]: I0309 09:53:07.823394 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" event={"ID":"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4","Type":"ContainerStarted","Data":"3741baa6a6599040021baf38e73307440f08f91d67d84947eb5332453d4ebe0b"} Mar 09 09:53:07 crc kubenswrapper[4692]: I0309 09:53:07.839938 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" podStartSLOduration=1.8399229689999999 podStartE2EDuration="1.839922969s" podCreationTimestamp="2026-03-09 09:53:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:07.838434396 +0000 UTC m=+1988.663170017" watchObservedRunningTime="2026-03-09 09:53:07.839922969 +0000 UTC m=+1988.664658550" Mar 09 09:53:08 crc kubenswrapper[4692]: I0309 09:53:08.833557 4692 generic.go:334] "Generic (PLEG): container finished" podID="1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" containerID="9770ea0091fa66c133afeb08906c31ff59c3bdc74f88c85deca0b7b375da682b" exitCode=0 Mar 09 09:53:08 crc kubenswrapper[4692]: I0309 09:53:08.833672 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" event={"ID":"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4","Type":"ContainerDied","Data":"9770ea0091fa66c133afeb08906c31ff59c3bdc74f88c85deca0b7b375da682b"} Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.072494 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.105470 4692 status_manager.go:875] "Failed to update status for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-09T09:53:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"}]}}\" for pod \"swift-kuttl-tests\"/\"swift-ring-rebalance-debug-pb77z\": pods \"swift-ring-rebalance-debug-pb77z\" not found" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.112863 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pb77z"] Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.118549 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pb77z"] Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.239069 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-scripts\") pod \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.239153 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-ring-data-devices\") pod \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.239193 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-swiftconf\") pod \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.239224 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz4hr\" (UniqueName: \"kubernetes.io/projected/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-kube-api-access-xz4hr\") pod \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.239319 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-dispersionconf\") pod \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.239448 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-etc-swift\") pod \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\" (UID: \"1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4\") " Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.239732 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" (UID: "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.240963 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" (UID: "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.245492 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-kube-api-access-xz4hr" (OuterVolumeSpecName: "kube-api-access-xz4hr") pod "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" (UID: "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4"). InnerVolumeSpecName "kube-api-access-xz4hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.260578 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-scripts" (OuterVolumeSpecName: "scripts") pod "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" (UID: "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.263976 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" (UID: "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.269378 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" (UID: "1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.342005 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.342071 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.342081 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.342091 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.342100 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz4hr\" (UniqueName: \"kubernetes.io/projected/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-kube-api-access-xz4hr\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.342108 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.849341 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3741baa6a6599040021baf38e73307440f08f91d67d84947eb5332453d4ebe0b" Mar 09 09:53:10 crc kubenswrapper[4692]: I0309 09:53:10.849409 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pb77z" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.252962 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622m4"] Mar 09 09:53:11 crc kubenswrapper[4692]: E0309 09:53:11.253282 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" containerName="swift-ring-rebalance" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.253300 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" containerName="swift-ring-rebalance" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.253475 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" containerName="swift-ring-rebalance" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.254059 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.257368 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.257470 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.267533 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622m4"] Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.355050 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-ring-data-devices\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.355113 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n84fw\" (UniqueName: \"kubernetes.io/projected/ce2f143c-ed98-4e31-bca8-f98f7854daf6-kube-api-access-n84fw\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.355149 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-scripts\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.355201 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ce2f143c-ed98-4e31-bca8-f98f7854daf6-etc-swift\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.355470 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-dispersionconf\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.355712 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-swiftconf\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457079 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-ring-data-devices\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457125 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n84fw\" (UniqueName: \"kubernetes.io/projected/ce2f143c-ed98-4e31-bca8-f98f7854daf6-kube-api-access-n84fw\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457145 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-scripts\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457213 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ce2f143c-ed98-4e31-bca8-f98f7854daf6-etc-swift\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457252 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-dispersionconf\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457282 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-swiftconf\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457951 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ce2f143c-ed98-4e31-bca8-f98f7854daf6-etc-swift\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.457994 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-ring-data-devices\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.458093 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-scripts\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.462765 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-swiftconf\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.470779 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-dispersionconf\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.477050 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n84fw\" (UniqueName: \"kubernetes.io/projected/ce2f143c-ed98-4e31-bca8-f98f7854daf6-kube-api-access-n84fw\") pod \"swift-ring-rebalance-debug-622m4\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.580986 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:11 crc kubenswrapper[4692]: I0309 09:53:11.981303 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622m4"] Mar 09 09:53:11 crc kubenswrapper[4692]: W0309 09:53:11.985378 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce2f143c_ed98_4e31_bca8_f98f7854daf6.slice/crio-cc8037a25ae15a188c367e054b5912d7b840b0dc6ddd1c1578b1a7584deb87f1 WatchSource:0}: Error finding container cc8037a25ae15a188c367e054b5912d7b840b0dc6ddd1c1578b1a7584deb87f1: Status 404 returned error can't find the container with id cc8037a25ae15a188c367e054b5912d7b840b0dc6ddd1c1578b1a7584deb87f1 Mar 09 09:53:12 crc kubenswrapper[4692]: I0309 09:53:12.081788 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4" path="/var/lib/kubelet/pods/1db8a5c4-955b-4a56-9827-f5d0b8f6fbf4/volumes" Mar 09 09:53:12 crc kubenswrapper[4692]: I0309 09:53:12.864784 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" event={"ID":"ce2f143c-ed98-4e31-bca8-f98f7854daf6","Type":"ContainerStarted","Data":"8a2e2289a00d1bb19f5449795b30f50a7cb8bb53977e98b02dad05c2006a8d74"} Mar 09 09:53:12 crc kubenswrapper[4692]: I0309 09:53:12.865084 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" event={"ID":"ce2f143c-ed98-4e31-bca8-f98f7854daf6","Type":"ContainerStarted","Data":"cc8037a25ae15a188c367e054b5912d7b840b0dc6ddd1c1578b1a7584deb87f1"} Mar 09 09:53:12 crc kubenswrapper[4692]: I0309 09:53:12.882584 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" podStartSLOduration=1.882564514 podStartE2EDuration="1.882564514s" podCreationTimestamp="2026-03-09 09:53:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:12.881007219 +0000 UTC m=+1993.705742800" watchObservedRunningTime="2026-03-09 09:53:12.882564514 +0000 UTC m=+1993.707300095" Mar 09 09:53:13 crc kubenswrapper[4692]: I0309 09:53:13.874495 4692 generic.go:334] "Generic (PLEG): container finished" podID="ce2f143c-ed98-4e31-bca8-f98f7854daf6" containerID="8a2e2289a00d1bb19f5449795b30f50a7cb8bb53977e98b02dad05c2006a8d74" exitCode=0 Mar 09 09:53:13 crc kubenswrapper[4692]: I0309 09:53:13.874602 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" event={"ID":"ce2f143c-ed98-4e31-bca8-f98f7854daf6","Type":"ContainerDied","Data":"8a2e2289a00d1bb19f5449795b30f50a7cb8bb53977e98b02dad05c2006a8d74"} Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.133854 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.165083 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622m4"] Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.173330 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622m4"] Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.211379 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-swiftconf\") pod \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.211446 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-ring-data-devices\") pod \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.211504 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-dispersionconf\") pod \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.211571 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n84fw\" (UniqueName: \"kubernetes.io/projected/ce2f143c-ed98-4e31-bca8-f98f7854daf6-kube-api-access-n84fw\") pod \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.211702 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ce2f143c-ed98-4e31-bca8-f98f7854daf6-etc-swift\") pod \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.211739 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-scripts\") pod \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\" (UID: \"ce2f143c-ed98-4e31-bca8-f98f7854daf6\") " Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.213111 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce2f143c-ed98-4e31-bca8-f98f7854daf6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ce2f143c-ed98-4e31-bca8-f98f7854daf6" (UID: "ce2f143c-ed98-4e31-bca8-f98f7854daf6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.214852 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ce2f143c-ed98-4e31-bca8-f98f7854daf6" (UID: "ce2f143c-ed98-4e31-bca8-f98f7854daf6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.219037 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce2f143c-ed98-4e31-bca8-f98f7854daf6-kube-api-access-n84fw" (OuterVolumeSpecName: "kube-api-access-n84fw") pod "ce2f143c-ed98-4e31-bca8-f98f7854daf6" (UID: "ce2f143c-ed98-4e31-bca8-f98f7854daf6"). InnerVolumeSpecName "kube-api-access-n84fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.236857 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-scripts" (OuterVolumeSpecName: "scripts") pod "ce2f143c-ed98-4e31-bca8-f98f7854daf6" (UID: "ce2f143c-ed98-4e31-bca8-f98f7854daf6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.238420 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ce2f143c-ed98-4e31-bca8-f98f7854daf6" (UID: "ce2f143c-ed98-4e31-bca8-f98f7854daf6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.238579 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ce2f143c-ed98-4e31-bca8-f98f7854daf6" (UID: "ce2f143c-ed98-4e31-bca8-f98f7854daf6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.313593 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ce2f143c-ed98-4e31-bca8-f98f7854daf6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.313687 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.313721 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.313737 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ce2f143c-ed98-4e31-bca8-f98f7854daf6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.313751 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ce2f143c-ed98-4e31-bca8-f98f7854daf6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.313762 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n84fw\" (UniqueName: \"kubernetes.io/projected/ce2f143c-ed98-4e31-bca8-f98f7854daf6-kube-api-access-n84fw\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.891042 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc8037a25ae15a188c367e054b5912d7b840b0dc6ddd1c1578b1a7584deb87f1" Mar 09 09:53:15 crc kubenswrapper[4692]: I0309 09:53:15.891109 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622m4" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.081218 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce2f143c-ed98-4e31-bca8-f98f7854daf6" path="/var/lib/kubelet/pods/ce2f143c-ed98-4e31-bca8-f98f7854daf6/volumes" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.336327 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m9497"] Mar 09 09:53:16 crc kubenswrapper[4692]: E0309 09:53:16.336730 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce2f143c-ed98-4e31-bca8-f98f7854daf6" containerName="swift-ring-rebalance" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.336762 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce2f143c-ed98-4e31-bca8-f98f7854daf6" containerName="swift-ring-rebalance" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.337008 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce2f143c-ed98-4e31-bca8-f98f7854daf6" containerName="swift-ring-rebalance" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.337463 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.339326 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.339972 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.347356 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m9497"] Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.428283 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-scripts\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.428680 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g45sj\" (UniqueName: \"kubernetes.io/projected/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-kube-api-access-g45sj\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.428715 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-etc-swift\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.428743 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-dispersionconf\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.428864 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-ring-data-devices\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.428891 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-swiftconf\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.530333 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-scripts\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.530429 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g45sj\" (UniqueName: \"kubernetes.io/projected/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-kube-api-access-g45sj\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.530455 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-etc-swift\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.530475 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-dispersionconf\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.530495 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-ring-data-devices\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.530513 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-swiftconf\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.531180 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-etc-swift\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.531493 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-ring-data-devices\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.531938 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-scripts\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.535286 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-dispersionconf\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.536059 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-swiftconf\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.549511 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g45sj\" (UniqueName: \"kubernetes.io/projected/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-kube-api-access-g45sj\") pod \"swift-ring-rebalance-debug-m9497\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:16 crc kubenswrapper[4692]: I0309 09:53:16.654592 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:17 crc kubenswrapper[4692]: I0309 09:53:17.076741 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m9497"] Mar 09 09:53:17 crc kubenswrapper[4692]: W0309 09:53:17.080893 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3ca2e34_c0f6_46df_b0ff_60f8a6145161.slice/crio-c08de7ad7f2523182102d6fceded05385677e9a66fe3e050661edd86a0dcb95b WatchSource:0}: Error finding container c08de7ad7f2523182102d6fceded05385677e9a66fe3e050661edd86a0dcb95b: Status 404 returned error can't find the container with id c08de7ad7f2523182102d6fceded05385677e9a66fe3e050661edd86a0dcb95b Mar 09 09:53:17 crc kubenswrapper[4692]: I0309 09:53:17.624821 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:53:17 crc kubenswrapper[4692]: I0309 09:53:17.624894 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:53:17 crc kubenswrapper[4692]: I0309 09:53:17.908943 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" event={"ID":"b3ca2e34-c0f6-46df-b0ff-60f8a6145161","Type":"ContainerStarted","Data":"6e235a46dd97fdddeb3c18a7b1b4798244db0a5765b2f7df05b044d79da2a1b6"} Mar 09 09:53:17 crc kubenswrapper[4692]: I0309 09:53:17.909531 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" event={"ID":"b3ca2e34-c0f6-46df-b0ff-60f8a6145161","Type":"ContainerStarted","Data":"c08de7ad7f2523182102d6fceded05385677e9a66fe3e050661edd86a0dcb95b"} Mar 09 09:53:17 crc kubenswrapper[4692]: I0309 09:53:17.932087 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" podStartSLOduration=1.9320621550000001 podStartE2EDuration="1.932062155s" podCreationTimestamp="2026-03-09 09:53:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:17.927106715 +0000 UTC m=+1998.751842316" watchObservedRunningTime="2026-03-09 09:53:17.932062155 +0000 UTC m=+1998.756797746" Mar 09 09:53:18 crc kubenswrapper[4692]: I0309 09:53:18.924436 4692 generic.go:334] "Generic (PLEG): container finished" podID="b3ca2e34-c0f6-46df-b0ff-60f8a6145161" containerID="6e235a46dd97fdddeb3c18a7b1b4798244db0a5765b2f7df05b044d79da2a1b6" exitCode=0 Mar 09 09:53:18 crc kubenswrapper[4692]: I0309 09:53:18.924486 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" event={"ID":"b3ca2e34-c0f6-46df-b0ff-60f8a6145161","Type":"ContainerDied","Data":"6e235a46dd97fdddeb3c18a7b1b4798244db0a5765b2f7df05b044d79da2a1b6"} Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.182636 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.216834 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m9497"] Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.222150 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-m9497"] Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.313250 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-scripts\") pod \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.313320 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g45sj\" (UniqueName: \"kubernetes.io/projected/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-kube-api-access-g45sj\") pod \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.313432 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-etc-swift\") pod \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.313460 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-ring-data-devices\") pod \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.313483 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-swiftconf\") pod \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.313508 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-dispersionconf\") pod \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\" (UID: \"b3ca2e34-c0f6-46df-b0ff-60f8a6145161\") " Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.314329 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b3ca2e34-c0f6-46df-b0ff-60f8a6145161" (UID: "b3ca2e34-c0f6-46df-b0ff-60f8a6145161"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.314706 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b3ca2e34-c0f6-46df-b0ff-60f8a6145161" (UID: "b3ca2e34-c0f6-46df-b0ff-60f8a6145161"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.318436 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-kube-api-access-g45sj" (OuterVolumeSpecName: "kube-api-access-g45sj") pod "b3ca2e34-c0f6-46df-b0ff-60f8a6145161" (UID: "b3ca2e34-c0f6-46df-b0ff-60f8a6145161"). InnerVolumeSpecName "kube-api-access-g45sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.333102 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-scripts" (OuterVolumeSpecName: "scripts") pod "b3ca2e34-c0f6-46df-b0ff-60f8a6145161" (UID: "b3ca2e34-c0f6-46df-b0ff-60f8a6145161"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.337760 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b3ca2e34-c0f6-46df-b0ff-60f8a6145161" (UID: "b3ca2e34-c0f6-46df-b0ff-60f8a6145161"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.339771 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b3ca2e34-c0f6-46df-b0ff-60f8a6145161" (UID: "b3ca2e34-c0f6-46df-b0ff-60f8a6145161"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.415366 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.415420 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g45sj\" (UniqueName: \"kubernetes.io/projected/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-kube-api-access-g45sj\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.415435 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.415448 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.415460 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.415472 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b3ca2e34-c0f6-46df-b0ff-60f8a6145161-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.940337 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c08de7ad7f2523182102d6fceded05385677e9a66fe3e050661edd86a0dcb95b" Mar 09 09:53:20 crc kubenswrapper[4692]: I0309 09:53:20.940420 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-m9497" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.354122 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm"] Mar 09 09:53:21 crc kubenswrapper[4692]: E0309 09:53:21.354430 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ca2e34-c0f6-46df-b0ff-60f8a6145161" containerName="swift-ring-rebalance" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.354447 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ca2e34-c0f6-46df-b0ff-60f8a6145161" containerName="swift-ring-rebalance" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.354618 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ca2e34-c0f6-46df-b0ff-60f8a6145161" containerName="swift-ring-rebalance" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.355151 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.357598 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.357901 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.381923 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm"] Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.530504 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/42c7b0e4-46a6-4b2c-b966-df166906675d-etc-swift\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.530562 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-scripts\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.530611 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-dispersionconf\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.530658 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5h6q\" (UniqueName: \"kubernetes.io/projected/42c7b0e4-46a6-4b2c-b966-df166906675d-kube-api-access-c5h6q\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.530698 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-swiftconf\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.530762 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-ring-data-devices\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632011 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-ring-data-devices\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632062 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/42c7b0e4-46a6-4b2c-b966-df166906675d-etc-swift\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632093 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-scripts\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632123 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-dispersionconf\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632158 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5h6q\" (UniqueName: \"kubernetes.io/projected/42c7b0e4-46a6-4b2c-b966-df166906675d-kube-api-access-c5h6q\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632194 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-swiftconf\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632599 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/42c7b0e4-46a6-4b2c-b966-df166906675d-etc-swift\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.632993 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-scripts\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.633138 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-ring-data-devices\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.636651 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-dispersionconf\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.644981 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-swiftconf\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.647894 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5h6q\" (UniqueName: \"kubernetes.io/projected/42c7b0e4-46a6-4b2c-b966-df166906675d-kube-api-access-c5h6q\") pod \"swift-ring-rebalance-debug-h4lnm\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.682814 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.934789 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm"] Mar 09 09:53:21 crc kubenswrapper[4692]: W0309 09:53:21.943837 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42c7b0e4_46a6_4b2c_b966_df166906675d.slice/crio-2c616836062a82e2d853a20dc55aa6ce394034ccc39ae71c612855836aa9aae4 WatchSource:0}: Error finding container 2c616836062a82e2d853a20dc55aa6ce394034ccc39ae71c612855836aa9aae4: Status 404 returned error can't find the container with id 2c616836062a82e2d853a20dc55aa6ce394034ccc39ae71c612855836aa9aae4 Mar 09 09:53:21 crc kubenswrapper[4692]: I0309 09:53:21.956972 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" event={"ID":"42c7b0e4-46a6-4b2c-b966-df166906675d","Type":"ContainerStarted","Data":"2c616836062a82e2d853a20dc55aa6ce394034ccc39ae71c612855836aa9aae4"} Mar 09 09:53:22 crc kubenswrapper[4692]: I0309 09:53:22.080119 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ca2e34-c0f6-46df-b0ff-60f8a6145161" path="/var/lib/kubelet/pods/b3ca2e34-c0f6-46df-b0ff-60f8a6145161/volumes" Mar 09 09:53:22 crc kubenswrapper[4692]: I0309 09:53:22.965529 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" event={"ID":"42c7b0e4-46a6-4b2c-b966-df166906675d","Type":"ContainerStarted","Data":"e5f0980ba2b99eca290cec57dc1e0c83f1bd0da6ffbcf57590d08929b8a48345"} Mar 09 09:53:22 crc kubenswrapper[4692]: I0309 09:53:22.987304 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" podStartSLOduration=1.987280498 podStartE2EDuration="1.987280498s" podCreationTimestamp="2026-03-09 09:53:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:22.98311322 +0000 UTC m=+2003.807848801" watchObservedRunningTime="2026-03-09 09:53:22.987280498 +0000 UTC m=+2003.812016079" Mar 09 09:53:23 crc kubenswrapper[4692]: I0309 09:53:23.974664 4692 generic.go:334] "Generic (PLEG): container finished" podID="42c7b0e4-46a6-4b2c-b966-df166906675d" containerID="e5f0980ba2b99eca290cec57dc1e0c83f1bd0da6ffbcf57590d08929b8a48345" exitCode=0 Mar 09 09:53:23 crc kubenswrapper[4692]: I0309 09:53:23.974792 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" event={"ID":"42c7b0e4-46a6-4b2c-b966-df166906675d","Type":"ContainerDied","Data":"e5f0980ba2b99eca290cec57dc1e0c83f1bd0da6ffbcf57590d08929b8a48345"} Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.248701 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.298263 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm"] Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.305766 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm"] Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.388468 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5h6q\" (UniqueName: \"kubernetes.io/projected/42c7b0e4-46a6-4b2c-b966-df166906675d-kube-api-access-c5h6q\") pod \"42c7b0e4-46a6-4b2c-b966-df166906675d\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.388567 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-scripts\") pod \"42c7b0e4-46a6-4b2c-b966-df166906675d\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.388595 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-swiftconf\") pod \"42c7b0e4-46a6-4b2c-b966-df166906675d\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.388648 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/42c7b0e4-46a6-4b2c-b966-df166906675d-etc-swift\") pod \"42c7b0e4-46a6-4b2c-b966-df166906675d\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.388677 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-ring-data-devices\") pod \"42c7b0e4-46a6-4b2c-b966-df166906675d\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.388715 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-dispersionconf\") pod \"42c7b0e4-46a6-4b2c-b966-df166906675d\" (UID: \"42c7b0e4-46a6-4b2c-b966-df166906675d\") " Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.389682 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "42c7b0e4-46a6-4b2c-b966-df166906675d" (UID: "42c7b0e4-46a6-4b2c-b966-df166906675d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.389988 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42c7b0e4-46a6-4b2c-b966-df166906675d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "42c7b0e4-46a6-4b2c-b966-df166906675d" (UID: "42c7b0e4-46a6-4b2c-b966-df166906675d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.411747 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42c7b0e4-46a6-4b2c-b966-df166906675d-kube-api-access-c5h6q" (OuterVolumeSpecName: "kube-api-access-c5h6q") pod "42c7b0e4-46a6-4b2c-b966-df166906675d" (UID: "42c7b0e4-46a6-4b2c-b966-df166906675d"). InnerVolumeSpecName "kube-api-access-c5h6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.434657 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-scripts" (OuterVolumeSpecName: "scripts") pod "42c7b0e4-46a6-4b2c-b966-df166906675d" (UID: "42c7b0e4-46a6-4b2c-b966-df166906675d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.455300 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "42c7b0e4-46a6-4b2c-b966-df166906675d" (UID: "42c7b0e4-46a6-4b2c-b966-df166906675d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.456568 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "42c7b0e4-46a6-4b2c-b966-df166906675d" (UID: "42c7b0e4-46a6-4b2c-b966-df166906675d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.490407 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5h6q\" (UniqueName: \"kubernetes.io/projected/42c7b0e4-46a6-4b2c-b966-df166906675d-kube-api-access-c5h6q\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.490456 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.490469 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.490478 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/42c7b0e4-46a6-4b2c-b966-df166906675d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.490488 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/42c7b0e4-46a6-4b2c-b966-df166906675d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.490498 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/42c7b0e4-46a6-4b2c-b966-df166906675d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.992609 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c616836062a82e2d853a20dc55aa6ce394034ccc39ae71c612855836aa9aae4" Mar 09 09:53:25 crc kubenswrapper[4692]: I0309 09:53:25.992679 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-h4lnm" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.081036 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42c7b0e4-46a6-4b2c-b966-df166906675d" path="/var/lib/kubelet/pods/42c7b0e4-46a6-4b2c-b966-df166906675d/volumes" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.428436 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb"] Mar 09 09:53:26 crc kubenswrapper[4692]: E0309 09:53:26.428879 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42c7b0e4-46a6-4b2c-b966-df166906675d" containerName="swift-ring-rebalance" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.428898 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="42c7b0e4-46a6-4b2c-b966-df166906675d" containerName="swift-ring-rebalance" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.429081 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="42c7b0e4-46a6-4b2c-b966-df166906675d" containerName="swift-ring-rebalance" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.429747 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.432296 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.432375 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.439188 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb"] Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.504329 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-ring-data-devices\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.504642 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-dispersionconf\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.504686 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30ecfe53-b59e-494c-bb20-cfb6f54849da-etc-swift\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.504714 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p968w\" (UniqueName: \"kubernetes.io/projected/30ecfe53-b59e-494c-bb20-cfb6f54849da-kube-api-access-p968w\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.504836 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-scripts\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.504877 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-swiftconf\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.606246 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-scripts\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.606312 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-swiftconf\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.606361 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-ring-data-devices\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.606390 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-dispersionconf\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.606426 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30ecfe53-b59e-494c-bb20-cfb6f54849da-etc-swift\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.606452 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p968w\" (UniqueName: \"kubernetes.io/projected/30ecfe53-b59e-494c-bb20-cfb6f54849da-kube-api-access-p968w\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.607092 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30ecfe53-b59e-494c-bb20-cfb6f54849da-etc-swift\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.607095 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-scripts\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.607124 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-ring-data-devices\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.611540 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-dispersionconf\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.613603 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-swiftconf\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.623571 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p968w\" (UniqueName: \"kubernetes.io/projected/30ecfe53-b59e-494c-bb20-cfb6f54849da-kube-api-access-p968w\") pod \"swift-ring-rebalance-debug-f8nkb\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:26 crc kubenswrapper[4692]: I0309 09:53:26.754709 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:27 crc kubenswrapper[4692]: I0309 09:53:27.197007 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb"] Mar 09 09:53:28 crc kubenswrapper[4692]: I0309 09:53:28.007611 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" event={"ID":"30ecfe53-b59e-494c-bb20-cfb6f54849da","Type":"ContainerStarted","Data":"869b74c8b7506c0252b5b0c7d192121a8c597138f6ebfd145a8875e9b4da5501"} Mar 09 09:53:28 crc kubenswrapper[4692]: I0309 09:53:28.007654 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" event={"ID":"30ecfe53-b59e-494c-bb20-cfb6f54849da","Type":"ContainerStarted","Data":"508d99a787a1b7b0bf3d53fcff220fe3ae30394e26396c9e9043dc4362058c9e"} Mar 09 09:53:29 crc kubenswrapper[4692]: I0309 09:53:29.016372 4692 generic.go:334] "Generic (PLEG): container finished" podID="30ecfe53-b59e-494c-bb20-cfb6f54849da" containerID="869b74c8b7506c0252b5b0c7d192121a8c597138f6ebfd145a8875e9b4da5501" exitCode=0 Mar 09 09:53:29 crc kubenswrapper[4692]: I0309 09:53:29.016433 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" event={"ID":"30ecfe53-b59e-494c-bb20-cfb6f54849da","Type":"ContainerDied","Data":"869b74c8b7506c0252b5b0c7d192121a8c597138f6ebfd145a8875e9b4da5501"} Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.315517 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.358602 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb"] Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.366210 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb"] Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.463222 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30ecfe53-b59e-494c-bb20-cfb6f54849da-etc-swift\") pod \"30ecfe53-b59e-494c-bb20-cfb6f54849da\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.463267 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-dispersionconf\") pod \"30ecfe53-b59e-494c-bb20-cfb6f54849da\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.463314 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-swiftconf\") pod \"30ecfe53-b59e-494c-bb20-cfb6f54849da\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.463384 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p968w\" (UniqueName: \"kubernetes.io/projected/30ecfe53-b59e-494c-bb20-cfb6f54849da-kube-api-access-p968w\") pod \"30ecfe53-b59e-494c-bb20-cfb6f54849da\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.463513 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-ring-data-devices\") pod \"30ecfe53-b59e-494c-bb20-cfb6f54849da\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.463540 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-scripts\") pod \"30ecfe53-b59e-494c-bb20-cfb6f54849da\" (UID: \"30ecfe53-b59e-494c-bb20-cfb6f54849da\") " Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.463984 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30ecfe53-b59e-494c-bb20-cfb6f54849da-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "30ecfe53-b59e-494c-bb20-cfb6f54849da" (UID: "30ecfe53-b59e-494c-bb20-cfb6f54849da"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.464089 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "30ecfe53-b59e-494c-bb20-cfb6f54849da" (UID: "30ecfe53-b59e-494c-bb20-cfb6f54849da"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.469470 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30ecfe53-b59e-494c-bb20-cfb6f54849da-kube-api-access-p968w" (OuterVolumeSpecName: "kube-api-access-p968w") pod "30ecfe53-b59e-494c-bb20-cfb6f54849da" (UID: "30ecfe53-b59e-494c-bb20-cfb6f54849da"). InnerVolumeSpecName "kube-api-access-p968w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.487998 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "30ecfe53-b59e-494c-bb20-cfb6f54849da" (UID: "30ecfe53-b59e-494c-bb20-cfb6f54849da"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.488204 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-scripts" (OuterVolumeSpecName: "scripts") pod "30ecfe53-b59e-494c-bb20-cfb6f54849da" (UID: "30ecfe53-b59e-494c-bb20-cfb6f54849da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.495387 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "30ecfe53-b59e-494c-bb20-cfb6f54849da" (UID: "30ecfe53-b59e-494c-bb20-cfb6f54849da"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.565080 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p968w\" (UniqueName: \"kubernetes.io/projected/30ecfe53-b59e-494c-bb20-cfb6f54849da-kube-api-access-p968w\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.565113 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.565124 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30ecfe53-b59e-494c-bb20-cfb6f54849da-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.565134 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30ecfe53-b59e-494c-bb20-cfb6f54849da-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.565143 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:30 crc kubenswrapper[4692]: I0309 09:53:30.565150 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30ecfe53-b59e-494c-bb20-cfb6f54849da-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.033454 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="508d99a787a1b7b0bf3d53fcff220fe3ae30394e26396c9e9043dc4362058c9e" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.033511 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f8nkb" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.506543 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2"] Mar 09 09:53:31 crc kubenswrapper[4692]: E0309 09:53:31.507117 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30ecfe53-b59e-494c-bb20-cfb6f54849da" containerName="swift-ring-rebalance" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.507133 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="30ecfe53-b59e-494c-bb20-cfb6f54849da" containerName="swift-ring-rebalance" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.507324 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="30ecfe53-b59e-494c-bb20-cfb6f54849da" containerName="swift-ring-rebalance" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.507973 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.509723 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.510893 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.513812 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2"] Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.679656 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-dispersionconf\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.681648 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4v9v\" (UniqueName: \"kubernetes.io/projected/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-kube-api-access-s4v9v\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.681835 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-scripts\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.681910 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-etc-swift\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.681959 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-ring-data-devices\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.681985 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-swiftconf\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.783774 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4v9v\" (UniqueName: \"kubernetes.io/projected/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-kube-api-access-s4v9v\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.783954 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-scripts\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.783999 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-etc-swift\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.784032 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-ring-data-devices\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.784057 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-swiftconf\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.784096 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-dispersionconf\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.784743 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-etc-swift\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.785424 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-scripts\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.785672 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-ring-data-devices\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.789579 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-swiftconf\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.795988 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-dispersionconf\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.809654 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4v9v\" (UniqueName: \"kubernetes.io/projected/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-kube-api-access-s4v9v\") pod \"swift-ring-rebalance-debug-ph4r2\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:31 crc kubenswrapper[4692]: I0309 09:53:31.833941 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:32 crc kubenswrapper[4692]: I0309 09:53:32.081093 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30ecfe53-b59e-494c-bb20-cfb6f54849da" path="/var/lib/kubelet/pods/30ecfe53-b59e-494c-bb20-cfb6f54849da/volumes" Mar 09 09:53:32 crc kubenswrapper[4692]: I0309 09:53:32.267686 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2"] Mar 09 09:53:32 crc kubenswrapper[4692]: W0309 09:53:32.275874 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecd7b586_931d_457d_89dd_9ea8a3a6d0ce.slice/crio-9b879b529dbbf59430e1c976df9add386b51d26cefc2e577f1e6153976f425a6 WatchSource:0}: Error finding container 9b879b529dbbf59430e1c976df9add386b51d26cefc2e577f1e6153976f425a6: Status 404 returned error can't find the container with id 9b879b529dbbf59430e1c976df9add386b51d26cefc2e577f1e6153976f425a6 Mar 09 09:53:33 crc kubenswrapper[4692]: I0309 09:53:33.050427 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" event={"ID":"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce","Type":"ContainerStarted","Data":"545ccfaec33d1b9164fefe6b86cd00cf58ef5ad54a99b5df0f4380a86f60b093"} Mar 09 09:53:33 crc kubenswrapper[4692]: I0309 09:53:33.050725 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" event={"ID":"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce","Type":"ContainerStarted","Data":"9b879b529dbbf59430e1c976df9add386b51d26cefc2e577f1e6153976f425a6"} Mar 09 09:53:33 crc kubenswrapper[4692]: I0309 09:53:33.068450 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" podStartSLOduration=2.068432823 podStartE2EDuration="2.068432823s" podCreationTimestamp="2026-03-09 09:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:33.066941631 +0000 UTC m=+2013.891677222" watchObservedRunningTime="2026-03-09 09:53:33.068432823 +0000 UTC m=+2013.893168404" Mar 09 09:53:34 crc kubenswrapper[4692]: I0309 09:53:34.058368 4692 generic.go:334] "Generic (PLEG): container finished" podID="ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" containerID="545ccfaec33d1b9164fefe6b86cd00cf58ef5ad54a99b5df0f4380a86f60b093" exitCode=0 Mar 09 09:53:34 crc kubenswrapper[4692]: I0309 09:53:34.058451 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" event={"ID":"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce","Type":"ContainerDied","Data":"545ccfaec33d1b9164fefe6b86cd00cf58ef5ad54a99b5df0f4380a86f60b093"} Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.385728 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.419373 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2"] Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.424074 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2"] Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.551490 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-scripts\") pod \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.551547 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-swiftconf\") pod \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.551592 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-etc-swift\") pod \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.551640 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-ring-data-devices\") pod \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.551700 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-dispersionconf\") pod \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.551738 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4v9v\" (UniqueName: \"kubernetes.io/projected/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-kube-api-access-s4v9v\") pod \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\" (UID: \"ecd7b586-931d-457d-89dd-9ea8a3a6d0ce\") " Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.552403 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" (UID: "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.552519 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" (UID: "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.561189 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-kube-api-access-s4v9v" (OuterVolumeSpecName: "kube-api-access-s4v9v") pod "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" (UID: "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce"). InnerVolumeSpecName "kube-api-access-s4v9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.574595 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-scripts" (OuterVolumeSpecName: "scripts") pod "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" (UID: "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.575944 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" (UID: "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.578135 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" (UID: "ecd7b586-931d-457d-89dd-9ea8a3a6d0ce"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.653252 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.653289 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.653300 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.653310 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.653327 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:35 crc kubenswrapper[4692]: I0309 09:53:35.653335 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4v9v\" (UniqueName: \"kubernetes.io/projected/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce-kube-api-access-s4v9v\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.079525 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" path="/var/lib/kubelet/pods/ecd7b586-931d-457d-89dd-9ea8a3a6d0ce/volumes" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.079731 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ph4r2" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.080050 4692 scope.go:117] "RemoveContainer" containerID="545ccfaec33d1b9164fefe6b86cd00cf58ef5ad54a99b5df0f4380a86f60b093" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.564306 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k"] Mar 09 09:53:36 crc kubenswrapper[4692]: E0309 09:53:36.564892 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" containerName="swift-ring-rebalance" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.564908 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" containerName="swift-ring-rebalance" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.565068 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd7b586-931d-457d-89dd-9ea8a3a6d0ce" containerName="swift-ring-rebalance" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.565586 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.567395 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.567429 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.577973 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k"] Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.666039 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-swiftconf\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.666087 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-scripts\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.666194 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e29a663-a2de-4e68-997f-cd83f52097f0-etc-swift\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.666221 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-ring-data-devices\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.666391 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-dispersionconf\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.666487 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9m8l\" (UniqueName: \"kubernetes.io/projected/3e29a663-a2de-4e68-997f-cd83f52097f0-kube-api-access-t9m8l\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.768238 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9m8l\" (UniqueName: \"kubernetes.io/projected/3e29a663-a2de-4e68-997f-cd83f52097f0-kube-api-access-t9m8l\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.768337 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-swiftconf\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.768375 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-scripts\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.768436 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e29a663-a2de-4e68-997f-cd83f52097f0-etc-swift\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.768464 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-ring-data-devices\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.768506 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-dispersionconf\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.769151 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e29a663-a2de-4e68-997f-cd83f52097f0-etc-swift\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.769456 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-ring-data-devices\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.769505 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-scripts\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.774890 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-dispersionconf\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.774912 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-swiftconf\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.787496 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9m8l\" (UniqueName: \"kubernetes.io/projected/3e29a663-a2de-4e68-997f-cd83f52097f0-kube-api-access-t9m8l\") pod \"swift-ring-rebalance-debug-kfj5k\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:36 crc kubenswrapper[4692]: I0309 09:53:36.882612 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:37 crc kubenswrapper[4692]: I0309 09:53:37.306365 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k"] Mar 09 09:53:37 crc kubenswrapper[4692]: W0309 09:53:37.322569 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e29a663_a2de_4e68_997f_cd83f52097f0.slice/crio-04782e642db258617a8b8ed951b69e23acc0047340e262a1d54f8b05c2df98ef WatchSource:0}: Error finding container 04782e642db258617a8b8ed951b69e23acc0047340e262a1d54f8b05c2df98ef: Status 404 returned error can't find the container with id 04782e642db258617a8b8ed951b69e23acc0047340e262a1d54f8b05c2df98ef Mar 09 09:53:38 crc kubenswrapper[4692]: I0309 09:53:38.101637 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" event={"ID":"3e29a663-a2de-4e68-997f-cd83f52097f0","Type":"ContainerStarted","Data":"649278ae134347f516ec669ac93194c96cb0eebb54fb43b495abcd13e1bf6de1"} Mar 09 09:53:38 crc kubenswrapper[4692]: I0309 09:53:38.101975 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" event={"ID":"3e29a663-a2de-4e68-997f-cd83f52097f0","Type":"ContainerStarted","Data":"04782e642db258617a8b8ed951b69e23acc0047340e262a1d54f8b05c2df98ef"} Mar 09 09:53:38 crc kubenswrapper[4692]: I0309 09:53:38.128392 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" podStartSLOduration=2.128365589 podStartE2EDuration="2.128365589s" podCreationTimestamp="2026-03-09 09:53:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:38.116995697 +0000 UTC m=+2018.941731278" watchObservedRunningTime="2026-03-09 09:53:38.128365589 +0000 UTC m=+2018.953101170" Mar 09 09:53:39 crc kubenswrapper[4692]: I0309 09:53:39.110743 4692 generic.go:334] "Generic (PLEG): container finished" podID="3e29a663-a2de-4e68-997f-cd83f52097f0" containerID="649278ae134347f516ec669ac93194c96cb0eebb54fb43b495abcd13e1bf6de1" exitCode=0 Mar 09 09:53:39 crc kubenswrapper[4692]: I0309 09:53:39.110797 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" event={"ID":"3e29a663-a2de-4e68-997f-cd83f52097f0","Type":"ContainerDied","Data":"649278ae134347f516ec669ac93194c96cb0eebb54fb43b495abcd13e1bf6de1"} Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.398559 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.428849 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k"] Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.434410 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k"] Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.524387 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-scripts\") pod \"3e29a663-a2de-4e68-997f-cd83f52097f0\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.524476 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-swiftconf\") pod \"3e29a663-a2de-4e68-997f-cd83f52097f0\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.524527 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-dispersionconf\") pod \"3e29a663-a2de-4e68-997f-cd83f52097f0\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.524556 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-ring-data-devices\") pod \"3e29a663-a2de-4e68-997f-cd83f52097f0\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.524687 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e29a663-a2de-4e68-997f-cd83f52097f0-etc-swift\") pod \"3e29a663-a2de-4e68-997f-cd83f52097f0\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.524710 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9m8l\" (UniqueName: \"kubernetes.io/projected/3e29a663-a2de-4e68-997f-cd83f52097f0-kube-api-access-t9m8l\") pod \"3e29a663-a2de-4e68-997f-cd83f52097f0\" (UID: \"3e29a663-a2de-4e68-997f-cd83f52097f0\") " Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.526596 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3e29a663-a2de-4e68-997f-cd83f52097f0" (UID: "3e29a663-a2de-4e68-997f-cd83f52097f0"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.526744 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e29a663-a2de-4e68-997f-cd83f52097f0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3e29a663-a2de-4e68-997f-cd83f52097f0" (UID: "3e29a663-a2de-4e68-997f-cd83f52097f0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.534721 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e29a663-a2de-4e68-997f-cd83f52097f0-kube-api-access-t9m8l" (OuterVolumeSpecName: "kube-api-access-t9m8l") pod "3e29a663-a2de-4e68-997f-cd83f52097f0" (UID: "3e29a663-a2de-4e68-997f-cd83f52097f0"). InnerVolumeSpecName "kube-api-access-t9m8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.545818 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-scripts" (OuterVolumeSpecName: "scripts") pod "3e29a663-a2de-4e68-997f-cd83f52097f0" (UID: "3e29a663-a2de-4e68-997f-cd83f52097f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.548077 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3e29a663-a2de-4e68-997f-cd83f52097f0" (UID: "3e29a663-a2de-4e68-997f-cd83f52097f0"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.549398 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3e29a663-a2de-4e68-997f-cd83f52097f0" (UID: "3e29a663-a2de-4e68-997f-cd83f52097f0"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.626142 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.626195 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.626207 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3e29a663-a2de-4e68-997f-cd83f52097f0-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.626215 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3e29a663-a2de-4e68-997f-cd83f52097f0-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.626226 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3e29a663-a2de-4e68-997f-cd83f52097f0-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:40 crc kubenswrapper[4692]: I0309 09:53:40.626234 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9m8l\" (UniqueName: \"kubernetes.io/projected/3e29a663-a2de-4e68-997f-cd83f52097f0-kube-api-access-t9m8l\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.129441 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04782e642db258617a8b8ed951b69e23acc0047340e262a1d54f8b05c2df98ef" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.129482 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kfj5k" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.569444 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf"] Mar 09 09:53:41 crc kubenswrapper[4692]: E0309 09:53:41.570028 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e29a663-a2de-4e68-997f-cd83f52097f0" containerName="swift-ring-rebalance" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.570045 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e29a663-a2de-4e68-997f-cd83f52097f0" containerName="swift-ring-rebalance" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.570285 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e29a663-a2de-4e68-997f-cd83f52097f0" containerName="swift-ring-rebalance" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.570848 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.572663 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.572851 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.585751 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf"] Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.642501 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-ring-data-devices\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.642593 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-swiftconf\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.642700 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-dispersionconf\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.642719 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-etc-swift\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.642765 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pzb7\" (UniqueName: \"kubernetes.io/projected/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-kube-api-access-5pzb7\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.642796 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-scripts\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.743760 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-scripts\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.743843 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-ring-data-devices\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.743903 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-swiftconf\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.743956 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-dispersionconf\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.743981 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-etc-swift\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.744030 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pzb7\" (UniqueName: \"kubernetes.io/projected/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-kube-api-access-5pzb7\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.744688 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-scripts\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.744979 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-ring-data-devices\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.745303 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-etc-swift\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.751813 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-swiftconf\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.752111 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-dispersionconf\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.766907 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pzb7\" (UniqueName: \"kubernetes.io/projected/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-kube-api-access-5pzb7\") pod \"swift-ring-rebalance-debug-9hnjf\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:41 crc kubenswrapper[4692]: I0309 09:53:41.887863 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:42 crc kubenswrapper[4692]: I0309 09:53:42.082527 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e29a663-a2de-4e68-997f-cd83f52097f0" path="/var/lib/kubelet/pods/3e29a663-a2de-4e68-997f-cd83f52097f0/volumes" Mar 09 09:53:42 crc kubenswrapper[4692]: I0309 09:53:42.336105 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf"] Mar 09 09:53:42 crc kubenswrapper[4692]: W0309 09:53:42.340599 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64c5afde_acc8_468a_a2bc_8d9121e1fa5c.slice/crio-2c9515cec7c04721a589e18ceb066c9f4e2e3d0b616a4233be5ff3e73251e92d WatchSource:0}: Error finding container 2c9515cec7c04721a589e18ceb066c9f4e2e3d0b616a4233be5ff3e73251e92d: Status 404 returned error can't find the container with id 2c9515cec7c04721a589e18ceb066c9f4e2e3d0b616a4233be5ff3e73251e92d Mar 09 09:53:43 crc kubenswrapper[4692]: I0309 09:53:43.148394 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" event={"ID":"64c5afde-acc8-468a-a2bc-8d9121e1fa5c","Type":"ContainerStarted","Data":"7dc328e88330d3655452f9867c65198a082b871ea8727254ae913c39b796de90"} Mar 09 09:53:43 crc kubenswrapper[4692]: I0309 09:53:43.148691 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" event={"ID":"64c5afde-acc8-468a-a2bc-8d9121e1fa5c","Type":"ContainerStarted","Data":"2c9515cec7c04721a589e18ceb066c9f4e2e3d0b616a4233be5ff3e73251e92d"} Mar 09 09:53:43 crc kubenswrapper[4692]: I0309 09:53:43.173012 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" podStartSLOduration=2.172987372 podStartE2EDuration="2.172987372s" podCreationTimestamp="2026-03-09 09:53:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:43.164481681 +0000 UTC m=+2023.989217282" watchObservedRunningTime="2026-03-09 09:53:43.172987372 +0000 UTC m=+2023.997722953" Mar 09 09:53:44 crc kubenswrapper[4692]: I0309 09:53:44.157606 4692 generic.go:334] "Generic (PLEG): container finished" podID="64c5afde-acc8-468a-a2bc-8d9121e1fa5c" containerID="7dc328e88330d3655452f9867c65198a082b871ea8727254ae913c39b796de90" exitCode=0 Mar 09 09:53:44 crc kubenswrapper[4692]: I0309 09:53:44.157661 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" event={"ID":"64c5afde-acc8-468a-a2bc-8d9121e1fa5c","Type":"ContainerDied","Data":"7dc328e88330d3655452f9867c65198a082b871ea8727254ae913c39b796de90"} Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.444153 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.485909 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf"] Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.496667 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf"] Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.595708 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pzb7\" (UniqueName: \"kubernetes.io/projected/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-kube-api-access-5pzb7\") pod \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.595851 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-swiftconf\") pod \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.595888 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-scripts\") pod \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.595917 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-dispersionconf\") pod \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.595942 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-ring-data-devices\") pod \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.596007 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-etc-swift\") pod \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\" (UID: \"64c5afde-acc8-468a-a2bc-8d9121e1fa5c\") " Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.596849 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "64c5afde-acc8-468a-a2bc-8d9121e1fa5c" (UID: "64c5afde-acc8-468a-a2bc-8d9121e1fa5c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.597497 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "64c5afde-acc8-468a-a2bc-8d9121e1fa5c" (UID: "64c5afde-acc8-468a-a2bc-8d9121e1fa5c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.601911 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-kube-api-access-5pzb7" (OuterVolumeSpecName: "kube-api-access-5pzb7") pod "64c5afde-acc8-468a-a2bc-8d9121e1fa5c" (UID: "64c5afde-acc8-468a-a2bc-8d9121e1fa5c"). InnerVolumeSpecName "kube-api-access-5pzb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.616143 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-scripts" (OuterVolumeSpecName: "scripts") pod "64c5afde-acc8-468a-a2bc-8d9121e1fa5c" (UID: "64c5afde-acc8-468a-a2bc-8d9121e1fa5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.623772 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "64c5afde-acc8-468a-a2bc-8d9121e1fa5c" (UID: "64c5afde-acc8-468a-a2bc-8d9121e1fa5c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.624749 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "64c5afde-acc8-468a-a2bc-8d9121e1fa5c" (UID: "64c5afde-acc8-468a-a2bc-8d9121e1fa5c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.697570 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.697611 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.697623 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.697633 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.697643 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pzb7\" (UniqueName: \"kubernetes.io/projected/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-kube-api-access-5pzb7\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:45 crc kubenswrapper[4692]: I0309 09:53:45.697654 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/64c5afde-acc8-468a-a2bc-8d9121e1fa5c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.082803 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64c5afde-acc8-468a-a2bc-8d9121e1fa5c" path="/var/lib/kubelet/pods/64c5afde-acc8-468a-a2bc-8d9121e1fa5c/volumes" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.171907 4692 scope.go:117] "RemoveContainer" containerID="7dc328e88330d3655452f9867c65198a082b871ea8727254ae913c39b796de90" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.171976 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-9hnjf" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.647290 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd"] Mar 09 09:53:46 crc kubenswrapper[4692]: E0309 09:53:46.647594 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c5afde-acc8-468a-a2bc-8d9121e1fa5c" containerName="swift-ring-rebalance" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.647606 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c5afde-acc8-468a-a2bc-8d9121e1fa5c" containerName="swift-ring-rebalance" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.647765 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c5afde-acc8-468a-a2bc-8d9121e1fa5c" containerName="swift-ring-rebalance" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.648311 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.657756 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd"] Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.658299 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.660759 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.711878 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g868n\" (UniqueName: \"kubernetes.io/projected/27d83ded-0aa9-450b-9ddb-e11e1628cef9-kube-api-access-g868n\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.711970 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-dispersionconf\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.712256 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-scripts\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.712342 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-ring-data-devices\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.712424 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27d83ded-0aa9-450b-9ddb-e11e1628cef9-etc-swift\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.712591 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-swiftconf\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.813695 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g868n\" (UniqueName: \"kubernetes.io/projected/27d83ded-0aa9-450b-9ddb-e11e1628cef9-kube-api-access-g868n\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.813768 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-dispersionconf\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.813814 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-scripts\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.813833 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-ring-data-devices\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.813854 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27d83ded-0aa9-450b-9ddb-e11e1628cef9-etc-swift\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.813909 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-swiftconf\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.814395 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27d83ded-0aa9-450b-9ddb-e11e1628cef9-etc-swift\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.814776 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-ring-data-devices\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.814870 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-scripts\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.818043 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-swiftconf\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.819749 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-dispersionconf\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.830879 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g868n\" (UniqueName: \"kubernetes.io/projected/27d83ded-0aa9-450b-9ddb-e11e1628cef9-kube-api-access-g868n\") pod \"swift-ring-rebalance-debug-rj6dd\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:46 crc kubenswrapper[4692]: I0309 09:53:46.971723 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:47 crc kubenswrapper[4692]: I0309 09:53:47.402182 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd"] Mar 09 09:53:47 crc kubenswrapper[4692]: I0309 09:53:47.624409 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:53:47 crc kubenswrapper[4692]: I0309 09:53:47.624468 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:53:48 crc kubenswrapper[4692]: I0309 09:53:48.193513 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" event={"ID":"27d83ded-0aa9-450b-9ddb-e11e1628cef9","Type":"ContainerStarted","Data":"44adc8a44c7bc291cc756cf5cb875bb165f1af67c97281123ade2b3897125a9c"} Mar 09 09:53:48 crc kubenswrapper[4692]: I0309 09:53:48.193558 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" event={"ID":"27d83ded-0aa9-450b-9ddb-e11e1628cef9","Type":"ContainerStarted","Data":"2df55c9a809f26e732c0ec03adfb882333dbb0d3a10fcc6d4785b92c5d505020"} Mar 09 09:53:48 crc kubenswrapper[4692]: I0309 09:53:48.208372 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" podStartSLOduration=2.2083531020000002 podStartE2EDuration="2.208353102s" podCreationTimestamp="2026-03-09 09:53:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:48.206744606 +0000 UTC m=+2029.031480197" watchObservedRunningTime="2026-03-09 09:53:48.208353102 +0000 UTC m=+2029.033088683" Mar 09 09:53:49 crc kubenswrapper[4692]: I0309 09:53:49.201410 4692 generic.go:334] "Generic (PLEG): container finished" podID="27d83ded-0aa9-450b-9ddb-e11e1628cef9" containerID="44adc8a44c7bc291cc756cf5cb875bb165f1af67c97281123ade2b3897125a9c" exitCode=0 Mar 09 09:53:49 crc kubenswrapper[4692]: I0309 09:53:49.201464 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" event={"ID":"27d83ded-0aa9-450b-9ddb-e11e1628cef9","Type":"ContainerDied","Data":"44adc8a44c7bc291cc756cf5cb875bb165f1af67c97281123ade2b3897125a9c"} Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.451343 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.485774 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd"] Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.491194 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd"] Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.562599 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-scripts\") pod \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.562917 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g868n\" (UniqueName: \"kubernetes.io/projected/27d83ded-0aa9-450b-9ddb-e11e1628cef9-kube-api-access-g868n\") pod \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.563026 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27d83ded-0aa9-450b-9ddb-e11e1628cef9-etc-swift\") pod \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.563119 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-swiftconf\") pod \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.563253 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-ring-data-devices\") pod \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.563329 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-dispersionconf\") pod \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\" (UID: \"27d83ded-0aa9-450b-9ddb-e11e1628cef9\") " Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.563767 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27d83ded-0aa9-450b-9ddb-e11e1628cef9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "27d83ded-0aa9-450b-9ddb-e11e1628cef9" (UID: "27d83ded-0aa9-450b-9ddb-e11e1628cef9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.564074 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "27d83ded-0aa9-450b-9ddb-e11e1628cef9" (UID: "27d83ded-0aa9-450b-9ddb-e11e1628cef9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.567962 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27d83ded-0aa9-450b-9ddb-e11e1628cef9-kube-api-access-g868n" (OuterVolumeSpecName: "kube-api-access-g868n") pod "27d83ded-0aa9-450b-9ddb-e11e1628cef9" (UID: "27d83ded-0aa9-450b-9ddb-e11e1628cef9"). InnerVolumeSpecName "kube-api-access-g868n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.583301 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-scripts" (OuterVolumeSpecName: "scripts") pod "27d83ded-0aa9-450b-9ddb-e11e1628cef9" (UID: "27d83ded-0aa9-450b-9ddb-e11e1628cef9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.589597 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "27d83ded-0aa9-450b-9ddb-e11e1628cef9" (UID: "27d83ded-0aa9-450b-9ddb-e11e1628cef9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.592517 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "27d83ded-0aa9-450b-9ddb-e11e1628cef9" (UID: "27d83ded-0aa9-450b-9ddb-e11e1628cef9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.666319 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.666359 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.666372 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g868n\" (UniqueName: \"kubernetes.io/projected/27d83ded-0aa9-450b-9ddb-e11e1628cef9-kube-api-access-g868n\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.666389 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27d83ded-0aa9-450b-9ddb-e11e1628cef9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.666401 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27d83ded-0aa9-450b-9ddb-e11e1628cef9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:50 crc kubenswrapper[4692]: I0309 09:53:50.666412 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27d83ded-0aa9-450b-9ddb-e11e1628cef9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.215409 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2df55c9a809f26e732c0ec03adfb882333dbb0d3a10fcc6d4785b92c5d505020" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.215475 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rj6dd" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.641486 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f"] Mar 09 09:53:51 crc kubenswrapper[4692]: E0309 09:53:51.641849 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27d83ded-0aa9-450b-9ddb-e11e1628cef9" containerName="swift-ring-rebalance" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.641865 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="27d83ded-0aa9-450b-9ddb-e11e1628cef9" containerName="swift-ring-rebalance" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.642053 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="27d83ded-0aa9-450b-9ddb-e11e1628cef9" containerName="swift-ring-rebalance" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.642661 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.650253 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.650299 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.669753 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f"] Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.781879 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-scripts\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.781922 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-swiftconf\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.781943 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwcfm\" (UniqueName: \"kubernetes.io/projected/bb16792c-60e9-499f-9ad2-75f422ecec94-kube-api-access-bwcfm\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.781964 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb16792c-60e9-499f-9ad2-75f422ecec94-etc-swift\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.781981 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-ring-data-devices\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.782408 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-dispersionconf\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.883916 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-scripts\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.883968 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-swiftconf\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.883989 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwcfm\" (UniqueName: \"kubernetes.io/projected/bb16792c-60e9-499f-9ad2-75f422ecec94-kube-api-access-bwcfm\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.884019 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb16792c-60e9-499f-9ad2-75f422ecec94-etc-swift\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.884043 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-ring-data-devices\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.884149 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-dispersionconf\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.884746 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb16792c-60e9-499f-9ad2-75f422ecec94-etc-swift\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.885026 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-ring-data-devices\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.885028 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-scripts\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.888580 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-dispersionconf\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.891954 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-swiftconf\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.904470 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwcfm\" (UniqueName: \"kubernetes.io/projected/bb16792c-60e9-499f-9ad2-75f422ecec94-kube-api-access-bwcfm\") pod \"swift-ring-rebalance-debug-jvc7f\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:51 crc kubenswrapper[4692]: I0309 09:53:51.961309 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:52 crc kubenswrapper[4692]: I0309 09:53:52.082349 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27d83ded-0aa9-450b-9ddb-e11e1628cef9" path="/var/lib/kubelet/pods/27d83ded-0aa9-450b-9ddb-e11e1628cef9/volumes" Mar 09 09:53:52 crc kubenswrapper[4692]: I0309 09:53:52.175597 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f"] Mar 09 09:53:52 crc kubenswrapper[4692]: W0309 09:53:52.178569 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb16792c_60e9_499f_9ad2_75f422ecec94.slice/crio-622bc0471e3e3898d95df4e79ad5f208423c316062c5b7c5f2c9cf8263301ef0 WatchSource:0}: Error finding container 622bc0471e3e3898d95df4e79ad5f208423c316062c5b7c5f2c9cf8263301ef0: Status 404 returned error can't find the container with id 622bc0471e3e3898d95df4e79ad5f208423c316062c5b7c5f2c9cf8263301ef0 Mar 09 09:53:52 crc kubenswrapper[4692]: I0309 09:53:52.231094 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" event={"ID":"bb16792c-60e9-499f-9ad2-75f422ecec94","Type":"ContainerStarted","Data":"622bc0471e3e3898d95df4e79ad5f208423c316062c5b7c5f2c9cf8263301ef0"} Mar 09 09:53:53 crc kubenswrapper[4692]: I0309 09:53:53.247951 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" event={"ID":"bb16792c-60e9-499f-9ad2-75f422ecec94","Type":"ContainerStarted","Data":"3a5de2c4293c96bf526ea6be6df1bf13012083bb62f7454b24c898e0b25a5451"} Mar 09 09:53:53 crc kubenswrapper[4692]: I0309 09:53:53.269478 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" podStartSLOduration=2.269459072 podStartE2EDuration="2.269459072s" podCreationTimestamp="2026-03-09 09:53:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:53.261065894 +0000 UTC m=+2034.085801495" watchObservedRunningTime="2026-03-09 09:53:53.269459072 +0000 UTC m=+2034.094194663" Mar 09 09:53:53 crc kubenswrapper[4692]: E0309 09:53:53.818299 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb16792c_60e9_499f_9ad2_75f422ecec94.slice/crio-conmon-3a5de2c4293c96bf526ea6be6df1bf13012083bb62f7454b24c898e0b25a5451.scope\": RecentStats: unable to find data in memory cache]" Mar 09 09:53:54 crc kubenswrapper[4692]: I0309 09:53:54.257129 4692 generic.go:334] "Generic (PLEG): container finished" podID="bb16792c-60e9-499f-9ad2-75f422ecec94" containerID="3a5de2c4293c96bf526ea6be6df1bf13012083bb62f7454b24c898e0b25a5451" exitCode=0 Mar 09 09:53:54 crc kubenswrapper[4692]: I0309 09:53:54.257182 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" event={"ID":"bb16792c-60e9-499f-9ad2-75f422ecec94","Type":"ContainerDied","Data":"3a5de2c4293c96bf526ea6be6df1bf13012083bb62f7454b24c898e0b25a5451"} Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.570248 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.600157 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f"] Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.607129 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f"] Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.658531 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-ring-data-devices\") pod \"bb16792c-60e9-499f-9ad2-75f422ecec94\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.659627 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb16792c-60e9-499f-9ad2-75f422ecec94-etc-swift\") pod \"bb16792c-60e9-499f-9ad2-75f422ecec94\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.659566 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bb16792c-60e9-499f-9ad2-75f422ecec94" (UID: "bb16792c-60e9-499f-9ad2-75f422ecec94"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.659713 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwcfm\" (UniqueName: \"kubernetes.io/projected/bb16792c-60e9-499f-9ad2-75f422ecec94-kube-api-access-bwcfm\") pod \"bb16792c-60e9-499f-9ad2-75f422ecec94\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.660451 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb16792c-60e9-499f-9ad2-75f422ecec94-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bb16792c-60e9-499f-9ad2-75f422ecec94" (UID: "bb16792c-60e9-499f-9ad2-75f422ecec94"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.660563 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-swiftconf\") pod \"bb16792c-60e9-499f-9ad2-75f422ecec94\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.660588 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-dispersionconf\") pod \"bb16792c-60e9-499f-9ad2-75f422ecec94\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.660989 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-scripts\") pod \"bb16792c-60e9-499f-9ad2-75f422ecec94\" (UID: \"bb16792c-60e9-499f-9ad2-75f422ecec94\") " Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.661472 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.661489 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bb16792c-60e9-499f-9ad2-75f422ecec94-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.667334 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb16792c-60e9-499f-9ad2-75f422ecec94-kube-api-access-bwcfm" (OuterVolumeSpecName: "kube-api-access-bwcfm") pod "bb16792c-60e9-499f-9ad2-75f422ecec94" (UID: "bb16792c-60e9-499f-9ad2-75f422ecec94"). InnerVolumeSpecName "kube-api-access-bwcfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.682139 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-scripts" (OuterVolumeSpecName: "scripts") pod "bb16792c-60e9-499f-9ad2-75f422ecec94" (UID: "bb16792c-60e9-499f-9ad2-75f422ecec94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.683050 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bb16792c-60e9-499f-9ad2-75f422ecec94" (UID: "bb16792c-60e9-499f-9ad2-75f422ecec94"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.684452 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bb16792c-60e9-499f-9ad2-75f422ecec94" (UID: "bb16792c-60e9-499f-9ad2-75f422ecec94"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.762678 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb16792c-60e9-499f-9ad2-75f422ecec94-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.762986 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwcfm\" (UniqueName: \"kubernetes.io/projected/bb16792c-60e9-499f-9ad2-75f422ecec94-kube-api-access-bwcfm\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.762997 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:55.763007 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bb16792c-60e9-499f-9ad2-75f422ecec94-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.081511 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb16792c-60e9-499f-9ad2-75f422ecec94" path="/var/lib/kubelet/pods/bb16792c-60e9-499f-9ad2-75f422ecec94/volumes" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.274329 4692 scope.go:117] "RemoveContainer" containerID="3a5de2c4293c96bf526ea6be6df1bf13012083bb62f7454b24c898e0b25a5451" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.274364 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jvc7f" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.751748 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q75tl"] Mar 09 09:53:56 crc kubenswrapper[4692]: E0309 09:53:56.752118 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb16792c-60e9-499f-9ad2-75f422ecec94" containerName="swift-ring-rebalance" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.752139 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb16792c-60e9-499f-9ad2-75f422ecec94" containerName="swift-ring-rebalance" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.752367 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb16792c-60e9-499f-9ad2-75f422ecec94" containerName="swift-ring-rebalance" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.752967 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.754889 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.754950 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.759651 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q75tl"] Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.880366 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-ring-data-devices\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.880473 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-swiftconf\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.880517 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-scripts\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.880537 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-dispersionconf\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.880581 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fcec7787-01f7-418d-a4d4-b33ac81e737a-etc-swift\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.880595 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgbxq\" (UniqueName: \"kubernetes.io/projected/fcec7787-01f7-418d-a4d4-b33ac81e737a-kube-api-access-kgbxq\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.982582 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-scripts\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.982651 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-dispersionconf\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.982751 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fcec7787-01f7-418d-a4d4-b33ac81e737a-etc-swift\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.982778 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgbxq\" (UniqueName: \"kubernetes.io/projected/fcec7787-01f7-418d-a4d4-b33ac81e737a-kube-api-access-kgbxq\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.982818 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-ring-data-devices\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.982890 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-swiftconf\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.983539 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fcec7787-01f7-418d-a4d4-b33ac81e737a-etc-swift\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.984008 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-ring-data-devices\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.984946 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-scripts\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.988810 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-dispersionconf\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:56 crc kubenswrapper[4692]: I0309 09:53:56.994793 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-swiftconf\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:57 crc kubenswrapper[4692]: I0309 09:53:57.001807 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgbxq\" (UniqueName: \"kubernetes.io/projected/fcec7787-01f7-418d-a4d4-b33ac81e737a-kube-api-access-kgbxq\") pod \"swift-ring-rebalance-debug-q75tl\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:57 crc kubenswrapper[4692]: I0309 09:53:57.069213 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:53:57 crc kubenswrapper[4692]: I0309 09:53:57.506068 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q75tl"] Mar 09 09:53:58 crc kubenswrapper[4692]: I0309 09:53:58.296819 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" event={"ID":"fcec7787-01f7-418d-a4d4-b33ac81e737a","Type":"ContainerStarted","Data":"8f2a2cc5ed500ff11c2c0e98a8a512d7cb04e99f8d914e177f065f4fc4fb99f4"} Mar 09 09:53:58 crc kubenswrapper[4692]: I0309 09:53:58.297108 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" event={"ID":"fcec7787-01f7-418d-a4d4-b33ac81e737a","Type":"ContainerStarted","Data":"e29ef24203ddca22221ec2a1c667756dde411b109dcd5f7f0af301f8601bec51"} Mar 09 09:53:58 crc kubenswrapper[4692]: I0309 09:53:58.324602 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" podStartSLOduration=2.324579402 podStartE2EDuration="2.324579402s" podCreationTimestamp="2026-03-09 09:53:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:53:58.315851075 +0000 UTC m=+2039.140586656" watchObservedRunningTime="2026-03-09 09:53:58.324579402 +0000 UTC m=+2039.149315003" Mar 09 09:53:59 crc kubenswrapper[4692]: I0309 09:53:59.309811 4692 generic.go:334] "Generic (PLEG): container finished" podID="fcec7787-01f7-418d-a4d4-b33ac81e737a" containerID="8f2a2cc5ed500ff11c2c0e98a8a512d7cb04e99f8d914e177f065f4fc4fb99f4" exitCode=0 Mar 09 09:53:59 crc kubenswrapper[4692]: I0309 09:53:59.309892 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" event={"ID":"fcec7787-01f7-418d-a4d4-b33ac81e737a","Type":"ContainerDied","Data":"8f2a2cc5ed500ff11c2c0e98a8a512d7cb04e99f8d914e177f065f4fc4fb99f4"} Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.136320 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550834-vbwpc"] Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.137804 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550834-vbwpc" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.141221 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.141711 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.142343 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.145895 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550834-vbwpc"] Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.232316 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4nl8\" (UniqueName: \"kubernetes.io/projected/5a0af45c-6d41-42fe-aa77-c587ed20b0dc-kube-api-access-f4nl8\") pod \"auto-csr-approver-29550834-vbwpc\" (UID: \"5a0af45c-6d41-42fe-aa77-c587ed20b0dc\") " pod="openshift-infra/auto-csr-approver-29550834-vbwpc" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.334336 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4nl8\" (UniqueName: \"kubernetes.io/projected/5a0af45c-6d41-42fe-aa77-c587ed20b0dc-kube-api-access-f4nl8\") pod \"auto-csr-approver-29550834-vbwpc\" (UID: \"5a0af45c-6d41-42fe-aa77-c587ed20b0dc\") " pod="openshift-infra/auto-csr-approver-29550834-vbwpc" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.358115 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4nl8\" (UniqueName: \"kubernetes.io/projected/5a0af45c-6d41-42fe-aa77-c587ed20b0dc-kube-api-access-f4nl8\") pod \"auto-csr-approver-29550834-vbwpc\" (UID: \"5a0af45c-6d41-42fe-aa77-c587ed20b0dc\") " pod="openshift-infra/auto-csr-approver-29550834-vbwpc" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.455780 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550834-vbwpc" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.676812 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.733016 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q75tl"] Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.741057 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q75tl"] Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.842145 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-scripts\") pod \"fcec7787-01f7-418d-a4d4-b33ac81e737a\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.842241 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-dispersionconf\") pod \"fcec7787-01f7-418d-a4d4-b33ac81e737a\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.842271 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-ring-data-devices\") pod \"fcec7787-01f7-418d-a4d4-b33ac81e737a\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.842288 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-swiftconf\") pod \"fcec7787-01f7-418d-a4d4-b33ac81e737a\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.842402 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgbxq\" (UniqueName: \"kubernetes.io/projected/fcec7787-01f7-418d-a4d4-b33ac81e737a-kube-api-access-kgbxq\") pod \"fcec7787-01f7-418d-a4d4-b33ac81e737a\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.842454 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fcec7787-01f7-418d-a4d4-b33ac81e737a-etc-swift\") pod \"fcec7787-01f7-418d-a4d4-b33ac81e737a\" (UID: \"fcec7787-01f7-418d-a4d4-b33ac81e737a\") " Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.843252 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fcec7787-01f7-418d-a4d4-b33ac81e737a" (UID: "fcec7787-01f7-418d-a4d4-b33ac81e737a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.843469 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcec7787-01f7-418d-a4d4-b33ac81e737a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fcec7787-01f7-418d-a4d4-b33ac81e737a" (UID: "fcec7787-01f7-418d-a4d4-b33ac81e737a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.847562 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcec7787-01f7-418d-a4d4-b33ac81e737a-kube-api-access-kgbxq" (OuterVolumeSpecName: "kube-api-access-kgbxq") pod "fcec7787-01f7-418d-a4d4-b33ac81e737a" (UID: "fcec7787-01f7-418d-a4d4-b33ac81e737a"). InnerVolumeSpecName "kube-api-access-kgbxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.862429 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fcec7787-01f7-418d-a4d4-b33ac81e737a" (UID: "fcec7787-01f7-418d-a4d4-b33ac81e737a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.862952 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-scripts" (OuterVolumeSpecName: "scripts") pod "fcec7787-01f7-418d-a4d4-b33ac81e737a" (UID: "fcec7787-01f7-418d-a4d4-b33ac81e737a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.869390 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fcec7787-01f7-418d-a4d4-b33ac81e737a" (UID: "fcec7787-01f7-418d-a4d4-b33ac81e737a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.894219 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550834-vbwpc"] Mar 09 09:54:00 crc kubenswrapper[4692]: W0309 09:54:00.897512 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a0af45c_6d41_42fe_aa77_c587ed20b0dc.slice/crio-81ef8b9959d316fc3d1dc4b1e715755ac74b235a2b7862142111291c5dec990b WatchSource:0}: Error finding container 81ef8b9959d316fc3d1dc4b1e715755ac74b235a2b7862142111291c5dec990b: Status 404 returned error can't find the container with id 81ef8b9959d316fc3d1dc4b1e715755ac74b235a2b7862142111291c5dec990b Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.944332 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.944365 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.944375 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fcec7787-01f7-418d-a4d4-b33ac81e737a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.944385 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgbxq\" (UniqueName: \"kubernetes.io/projected/fcec7787-01f7-418d-a4d4-b33ac81e737a-kube-api-access-kgbxq\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.944394 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fcec7787-01f7-418d-a4d4-b33ac81e737a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:00 crc kubenswrapper[4692]: I0309 09:54:00.944402 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcec7787-01f7-418d-a4d4-b33ac81e737a-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.324281 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550834-vbwpc" event={"ID":"5a0af45c-6d41-42fe-aa77-c587ed20b0dc","Type":"ContainerStarted","Data":"81ef8b9959d316fc3d1dc4b1e715755ac74b235a2b7862142111291c5dec990b"} Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.325865 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e29ef24203ddca22221ec2a1c667756dde411b109dcd5f7f0af301f8601bec51" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.325917 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q75tl" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.861743 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx"] Mar 09 09:54:01 crc kubenswrapper[4692]: E0309 09:54:01.862508 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcec7787-01f7-418d-a4d4-b33ac81e737a" containerName="swift-ring-rebalance" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.862524 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcec7787-01f7-418d-a4d4-b33ac81e737a" containerName="swift-ring-rebalance" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.862703 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcec7787-01f7-418d-a4d4-b33ac81e737a" containerName="swift-ring-rebalance" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.863934 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.865778 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.866117 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.870601 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx"] Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.958821 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-scripts\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.959028 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-dispersionconf\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.959102 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f8cg\" (UniqueName: \"kubernetes.io/projected/4c45e196-efb6-41f3-9a8a-c6ca4930674b-kube-api-access-9f8cg\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.959154 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c45e196-efb6-41f3-9a8a-c6ca4930674b-etc-swift\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.959239 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-swiftconf\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:01 crc kubenswrapper[4692]: I0309 09:54:01.959277 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-ring-data-devices\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.060499 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f8cg\" (UniqueName: \"kubernetes.io/projected/4c45e196-efb6-41f3-9a8a-c6ca4930674b-kube-api-access-9f8cg\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.060572 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c45e196-efb6-41f3-9a8a-c6ca4930674b-etc-swift\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.060609 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-swiftconf\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.060632 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-ring-data-devices\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.060658 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-scripts\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.060726 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-dispersionconf\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.061455 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c45e196-efb6-41f3-9a8a-c6ca4930674b-etc-swift\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.062294 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-scripts\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.062800 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-ring-data-devices\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.066153 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-swiftconf\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.072725 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-dispersionconf\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.081143 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f8cg\" (UniqueName: \"kubernetes.io/projected/4c45e196-efb6-41f3-9a8a-c6ca4930674b-kube-api-access-9f8cg\") pod \"swift-ring-rebalance-debug-dxgsx\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.081636 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcec7787-01f7-418d-a4d4-b33ac81e737a" path="/var/lib/kubelet/pods/fcec7787-01f7-418d-a4d4-b33ac81e737a/volumes" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.186478 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:02 crc kubenswrapper[4692]: I0309 09:54:02.598645 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx"] Mar 09 09:54:02 crc kubenswrapper[4692]: W0309 09:54:02.606437 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c45e196_efb6_41f3_9a8a_c6ca4930674b.slice/crio-2cf43c869ef6aed32a9a215d84f27ba3e3f1e6798b545a20166118a8d69378d9 WatchSource:0}: Error finding container 2cf43c869ef6aed32a9a215d84f27ba3e3f1e6798b545a20166118a8d69378d9: Status 404 returned error can't find the container with id 2cf43c869ef6aed32a9a215d84f27ba3e3f1e6798b545a20166118a8d69378d9 Mar 09 09:54:03 crc kubenswrapper[4692]: I0309 09:54:03.343217 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" event={"ID":"4c45e196-efb6-41f3-9a8a-c6ca4930674b","Type":"ContainerStarted","Data":"a8e80602f366b8ca14ce65d308157ab0bd34e9ba5f7edf4509dad3ebce4d060f"} Mar 09 09:54:03 crc kubenswrapper[4692]: I0309 09:54:03.343556 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" event={"ID":"4c45e196-efb6-41f3-9a8a-c6ca4930674b","Type":"ContainerStarted","Data":"2cf43c869ef6aed32a9a215d84f27ba3e3f1e6798b545a20166118a8d69378d9"} Mar 09 09:54:03 crc kubenswrapper[4692]: I0309 09:54:03.346025 4692 generic.go:334] "Generic (PLEG): container finished" podID="5a0af45c-6d41-42fe-aa77-c587ed20b0dc" containerID="43f05d768a18ca113aa566c302d271df491fefed968e2ee4a9f5257cbb1029cb" exitCode=0 Mar 09 09:54:03 crc kubenswrapper[4692]: I0309 09:54:03.346075 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550834-vbwpc" event={"ID":"5a0af45c-6d41-42fe-aa77-c587ed20b0dc","Type":"ContainerDied","Data":"43f05d768a18ca113aa566c302d271df491fefed968e2ee4a9f5257cbb1029cb"} Mar 09 09:54:03 crc kubenswrapper[4692]: I0309 09:54:03.364737 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" podStartSLOduration=2.364721877 podStartE2EDuration="2.364721877s" podCreationTimestamp="2026-03-09 09:54:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:03.360609701 +0000 UTC m=+2044.185345282" watchObservedRunningTime="2026-03-09 09:54:03.364721877 +0000 UTC m=+2044.189457458" Mar 09 09:54:04 crc kubenswrapper[4692]: I0309 09:54:04.355898 4692 generic.go:334] "Generic (PLEG): container finished" podID="4c45e196-efb6-41f3-9a8a-c6ca4930674b" containerID="a8e80602f366b8ca14ce65d308157ab0bd34e9ba5f7edf4509dad3ebce4d060f" exitCode=0 Mar 09 09:54:04 crc kubenswrapper[4692]: I0309 09:54:04.356343 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" event={"ID":"4c45e196-efb6-41f3-9a8a-c6ca4930674b","Type":"ContainerDied","Data":"a8e80602f366b8ca14ce65d308157ab0bd34e9ba5f7edf4509dad3ebce4d060f"} Mar 09 09:54:04 crc kubenswrapper[4692]: I0309 09:54:04.629964 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550834-vbwpc" Mar 09 09:54:04 crc kubenswrapper[4692]: I0309 09:54:04.693611 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4nl8\" (UniqueName: \"kubernetes.io/projected/5a0af45c-6d41-42fe-aa77-c587ed20b0dc-kube-api-access-f4nl8\") pod \"5a0af45c-6d41-42fe-aa77-c587ed20b0dc\" (UID: \"5a0af45c-6d41-42fe-aa77-c587ed20b0dc\") " Mar 09 09:54:04 crc kubenswrapper[4692]: I0309 09:54:04.699089 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a0af45c-6d41-42fe-aa77-c587ed20b0dc-kube-api-access-f4nl8" (OuterVolumeSpecName: "kube-api-access-f4nl8") pod "5a0af45c-6d41-42fe-aa77-c587ed20b0dc" (UID: "5a0af45c-6d41-42fe-aa77-c587ed20b0dc"). InnerVolumeSpecName "kube-api-access-f4nl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:04 crc kubenswrapper[4692]: I0309 09:54:04.794830 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4nl8\" (UniqueName: \"kubernetes.io/projected/5a0af45c-6d41-42fe-aa77-c587ed20b0dc-kube-api-access-f4nl8\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.364929 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550834-vbwpc" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.364938 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550834-vbwpc" event={"ID":"5a0af45c-6d41-42fe-aa77-c587ed20b0dc","Type":"ContainerDied","Data":"81ef8b9959d316fc3d1dc4b1e715755ac74b235a2b7862142111291c5dec990b"} Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.365299 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81ef8b9959d316fc3d1dc4b1e715755ac74b235a2b7862142111291c5dec990b" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.640910 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.677990 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx"] Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.683850 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx"] Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.702108 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550828-dthsk"] Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.707801 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550828-dthsk"] Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.708245 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-dispersionconf\") pod \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.708293 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f8cg\" (UniqueName: \"kubernetes.io/projected/4c45e196-efb6-41f3-9a8a-c6ca4930674b-kube-api-access-9f8cg\") pod \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.708368 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-scripts\") pod \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.708437 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-swiftconf\") pod \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.708454 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-ring-data-devices\") pod \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.708497 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c45e196-efb6-41f3-9a8a-c6ca4930674b-etc-swift\") pod \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\" (UID: \"4c45e196-efb6-41f3-9a8a-c6ca4930674b\") " Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.709048 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4c45e196-efb6-41f3-9a8a-c6ca4930674b" (UID: "4c45e196-efb6-41f3-9a8a-c6ca4930674b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.709274 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c45e196-efb6-41f3-9a8a-c6ca4930674b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4c45e196-efb6-41f3-9a8a-c6ca4930674b" (UID: "4c45e196-efb6-41f3-9a8a-c6ca4930674b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.719374 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c45e196-efb6-41f3-9a8a-c6ca4930674b-kube-api-access-9f8cg" (OuterVolumeSpecName: "kube-api-access-9f8cg") pod "4c45e196-efb6-41f3-9a8a-c6ca4930674b" (UID: "4c45e196-efb6-41f3-9a8a-c6ca4930674b"). InnerVolumeSpecName "kube-api-access-9f8cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.730347 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4c45e196-efb6-41f3-9a8a-c6ca4930674b" (UID: "4c45e196-efb6-41f3-9a8a-c6ca4930674b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.732141 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-scripts" (OuterVolumeSpecName: "scripts") pod "4c45e196-efb6-41f3-9a8a-c6ca4930674b" (UID: "4c45e196-efb6-41f3-9a8a-c6ca4930674b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.735207 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4c45e196-efb6-41f3-9a8a-c6ca4930674b" (UID: "4c45e196-efb6-41f3-9a8a-c6ca4930674b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.810528 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.810565 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.810578 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c45e196-efb6-41f3-9a8a-c6ca4930674b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.810594 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c45e196-efb6-41f3-9a8a-c6ca4930674b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.810604 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c45e196-efb6-41f3-9a8a-c6ca4930674b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:05 crc kubenswrapper[4692]: I0309 09:54:05.810618 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f8cg\" (UniqueName: \"kubernetes.io/projected/4c45e196-efb6-41f3-9a8a-c6ca4930674b-kube-api-access-9f8cg\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.080772 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd6688d-ce93-4f18-8b81-17f28509fe24" path="/var/lib/kubelet/pods/3dd6688d-ce93-4f18-8b81-17f28509fe24/volumes" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.081574 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c45e196-efb6-41f3-9a8a-c6ca4930674b" path="/var/lib/kubelet/pods/4c45e196-efb6-41f3-9a8a-c6ca4930674b/volumes" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.375868 4692 scope.go:117] "RemoveContainer" containerID="a8e80602f366b8ca14ce65d308157ab0bd34e9ba5f7edf4509dad3ebce4d060f" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.375880 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-dxgsx" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.830446 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76s56"] Mar 09 09:54:06 crc kubenswrapper[4692]: E0309 09:54:06.830770 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a0af45c-6d41-42fe-aa77-c587ed20b0dc" containerName="oc" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.830792 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a0af45c-6d41-42fe-aa77-c587ed20b0dc" containerName="oc" Mar 09 09:54:06 crc kubenswrapper[4692]: E0309 09:54:06.830819 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c45e196-efb6-41f3-9a8a-c6ca4930674b" containerName="swift-ring-rebalance" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.830831 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c45e196-efb6-41f3-9a8a-c6ca4930674b" containerName="swift-ring-rebalance" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.830961 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c45e196-efb6-41f3-9a8a-c6ca4930674b" containerName="swift-ring-rebalance" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.830988 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a0af45c-6d41-42fe-aa77-c587ed20b0dc" containerName="oc" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.831556 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.843803 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76s56"] Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.844231 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.844275 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.930966 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-scripts\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.931029 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-dispersionconf\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.931104 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-ring-data-devices\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.931183 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/045956af-b079-42ed-a4e6-042f1fe6a3dc-etc-swift\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.931260 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47wdg\" (UniqueName: \"kubernetes.io/projected/045956af-b079-42ed-a4e6-042f1fe6a3dc-kube-api-access-47wdg\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:06 crc kubenswrapper[4692]: I0309 09:54:06.931293 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-swiftconf\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.032552 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/045956af-b079-42ed-a4e6-042f1fe6a3dc-etc-swift\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.032674 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47wdg\" (UniqueName: \"kubernetes.io/projected/045956af-b079-42ed-a4e6-042f1fe6a3dc-kube-api-access-47wdg\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.032714 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-swiftconf\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.032754 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-scripts\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.032784 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-dispersionconf\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.032806 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-ring-data-devices\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.033124 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/045956af-b079-42ed-a4e6-042f1fe6a3dc-etc-swift\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.033825 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-scripts\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.033829 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-ring-data-devices\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.042513 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-swiftconf\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.043523 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-dispersionconf\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.051053 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47wdg\" (UniqueName: \"kubernetes.io/projected/045956af-b079-42ed-a4e6-042f1fe6a3dc-kube-api-access-47wdg\") pod \"swift-ring-rebalance-debug-76s56\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.161890 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:07 crc kubenswrapper[4692]: I0309 09:54:07.585207 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76s56"] Mar 09 09:54:07 crc kubenswrapper[4692]: W0309 09:54:07.597406 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod045956af_b079_42ed_a4e6_042f1fe6a3dc.slice/crio-c7867882152a342f6494d50732ee57eb0bbe22fd582082677fa921e9c5483feb WatchSource:0}: Error finding container c7867882152a342f6494d50732ee57eb0bbe22fd582082677fa921e9c5483feb: Status 404 returned error can't find the container with id c7867882152a342f6494d50732ee57eb0bbe22fd582082677fa921e9c5483feb Mar 09 09:54:08 crc kubenswrapper[4692]: I0309 09:54:08.394522 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" event={"ID":"045956af-b079-42ed-a4e6-042f1fe6a3dc","Type":"ContainerStarted","Data":"0870e7d45a9beb4862343e1a3b02cf223c1aa8ab6a20e1625b1d78e3458bf5de"} Mar 09 09:54:08 crc kubenswrapper[4692]: I0309 09:54:08.394870 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" event={"ID":"045956af-b079-42ed-a4e6-042f1fe6a3dc","Type":"ContainerStarted","Data":"c7867882152a342f6494d50732ee57eb0bbe22fd582082677fa921e9c5483feb"} Mar 09 09:54:08 crc kubenswrapper[4692]: I0309 09:54:08.416177 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" podStartSLOduration=2.416148733 podStartE2EDuration="2.416148733s" podCreationTimestamp="2026-03-09 09:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:08.40756285 +0000 UTC m=+2049.232298441" watchObservedRunningTime="2026-03-09 09:54:08.416148733 +0000 UTC m=+2049.240884304" Mar 09 09:54:09 crc kubenswrapper[4692]: I0309 09:54:09.402080 4692 generic.go:334] "Generic (PLEG): container finished" podID="045956af-b079-42ed-a4e6-042f1fe6a3dc" containerID="0870e7d45a9beb4862343e1a3b02cf223c1aa8ab6a20e1625b1d78e3458bf5de" exitCode=0 Mar 09 09:54:09 crc kubenswrapper[4692]: I0309 09:54:09.402125 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" event={"ID":"045956af-b079-42ed-a4e6-042f1fe6a3dc","Type":"ContainerDied","Data":"0870e7d45a9beb4862343e1a3b02cf223c1aa8ab6a20e1625b1d78e3458bf5de"} Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.667835 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.699427 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76s56"] Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.710435 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-76s56"] Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.785804 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-scripts\") pod \"045956af-b079-42ed-a4e6-042f1fe6a3dc\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.785883 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-ring-data-devices\") pod \"045956af-b079-42ed-a4e6-042f1fe6a3dc\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.785926 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/045956af-b079-42ed-a4e6-042f1fe6a3dc-etc-swift\") pod \"045956af-b079-42ed-a4e6-042f1fe6a3dc\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.786046 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-swiftconf\") pod \"045956af-b079-42ed-a4e6-042f1fe6a3dc\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.786086 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47wdg\" (UniqueName: \"kubernetes.io/projected/045956af-b079-42ed-a4e6-042f1fe6a3dc-kube-api-access-47wdg\") pod \"045956af-b079-42ed-a4e6-042f1fe6a3dc\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.786120 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-dispersionconf\") pod \"045956af-b079-42ed-a4e6-042f1fe6a3dc\" (UID: \"045956af-b079-42ed-a4e6-042f1fe6a3dc\") " Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.786699 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "045956af-b079-42ed-a4e6-042f1fe6a3dc" (UID: "045956af-b079-42ed-a4e6-042f1fe6a3dc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.786752 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/045956af-b079-42ed-a4e6-042f1fe6a3dc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "045956af-b079-42ed-a4e6-042f1fe6a3dc" (UID: "045956af-b079-42ed-a4e6-042f1fe6a3dc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.791586 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/045956af-b079-42ed-a4e6-042f1fe6a3dc-kube-api-access-47wdg" (OuterVolumeSpecName: "kube-api-access-47wdg") pod "045956af-b079-42ed-a4e6-042f1fe6a3dc" (UID: "045956af-b079-42ed-a4e6-042f1fe6a3dc"). InnerVolumeSpecName "kube-api-access-47wdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.810180 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "045956af-b079-42ed-a4e6-042f1fe6a3dc" (UID: "045956af-b079-42ed-a4e6-042f1fe6a3dc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.810692 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-scripts" (OuterVolumeSpecName: "scripts") pod "045956af-b079-42ed-a4e6-042f1fe6a3dc" (UID: "045956af-b079-42ed-a4e6-042f1fe6a3dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.810705 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "045956af-b079-42ed-a4e6-042f1fe6a3dc" (UID: "045956af-b079-42ed-a4e6-042f1fe6a3dc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.893342 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.893384 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/045956af-b079-42ed-a4e6-042f1fe6a3dc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.893401 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/045956af-b079-42ed-a4e6-042f1fe6a3dc-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.893415 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.893428 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47wdg\" (UniqueName: \"kubernetes.io/projected/045956af-b079-42ed-a4e6-042f1fe6a3dc-kube-api-access-47wdg\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:10 crc kubenswrapper[4692]: I0309 09:54:10.893440 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/045956af-b079-42ed-a4e6-042f1fe6a3dc-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.430411 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7867882152a342f6494d50732ee57eb0bbe22fd582082677fa921e9c5483feb" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.430560 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-76s56" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.842218 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4"] Mar 09 09:54:11 crc kubenswrapper[4692]: E0309 09:54:11.842515 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="045956af-b079-42ed-a4e6-042f1fe6a3dc" containerName="swift-ring-rebalance" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.842529 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="045956af-b079-42ed-a4e6-042f1fe6a3dc" containerName="swift-ring-rebalance" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.842707 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="045956af-b079-42ed-a4e6-042f1fe6a3dc" containerName="swift-ring-rebalance" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.843141 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.845315 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.850477 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.850746 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4"] Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.907812 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-dispersionconf\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.908690 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjmlb\" (UniqueName: \"kubernetes.io/projected/bd5610b5-0c3c-4673-9400-b571f4d07531-kube-api-access-bjmlb\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.908754 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-swiftconf\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.908847 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-scripts\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.909498 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-ring-data-devices\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:11 crc kubenswrapper[4692]: I0309 09:54:11.909576 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd5610b5-0c3c-4673-9400-b571f4d07531-etc-swift\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.011504 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-scripts\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.011570 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-ring-data-devices\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.011600 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd5610b5-0c3c-4673-9400-b571f4d07531-etc-swift\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.011650 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-dispersionconf\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.011682 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjmlb\" (UniqueName: \"kubernetes.io/projected/bd5610b5-0c3c-4673-9400-b571f4d07531-kube-api-access-bjmlb\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.011723 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-swiftconf\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.013074 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd5610b5-0c3c-4673-9400-b571f4d07531-etc-swift\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.013446 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-ring-data-devices\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.013459 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-scripts\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.017475 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-dispersionconf\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.024722 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-swiftconf\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.033950 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjmlb\" (UniqueName: \"kubernetes.io/projected/bd5610b5-0c3c-4673-9400-b571f4d07531-kube-api-access-bjmlb\") pod \"swift-ring-rebalance-debug-2c9z4\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.086218 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="045956af-b079-42ed-a4e6-042f1fe6a3dc" path="/var/lib/kubelet/pods/045956af-b079-42ed-a4e6-042f1fe6a3dc/volumes" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.158312 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:12 crc kubenswrapper[4692]: I0309 09:54:12.573697 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4"] Mar 09 09:54:12 crc kubenswrapper[4692]: W0309 09:54:12.579088 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd5610b5_0c3c_4673_9400_b571f4d07531.slice/crio-834de8bd8ec51c3a566b214cae255902f7f8a2e6a6ab92c4f4efcddb66c2c60e WatchSource:0}: Error finding container 834de8bd8ec51c3a566b214cae255902f7f8a2e6a6ab92c4f4efcddb66c2c60e: Status 404 returned error can't find the container with id 834de8bd8ec51c3a566b214cae255902f7f8a2e6a6ab92c4f4efcddb66c2c60e Mar 09 09:54:13 crc kubenswrapper[4692]: I0309 09:54:13.447469 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" event={"ID":"bd5610b5-0c3c-4673-9400-b571f4d07531","Type":"ContainerStarted","Data":"33dd9b6ec14546fb13067eb98fd438f8a369face0c9c0525fee2717b02c98869"} Mar 09 09:54:13 crc kubenswrapper[4692]: I0309 09:54:13.447527 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" event={"ID":"bd5610b5-0c3c-4673-9400-b571f4d07531","Type":"ContainerStarted","Data":"834de8bd8ec51c3a566b214cae255902f7f8a2e6a6ab92c4f4efcddb66c2c60e"} Mar 09 09:54:13 crc kubenswrapper[4692]: I0309 09:54:13.465113 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" podStartSLOduration=2.465096037 podStartE2EDuration="2.465096037s" podCreationTimestamp="2026-03-09 09:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:13.463853462 +0000 UTC m=+2054.288589043" watchObservedRunningTime="2026-03-09 09:54:13.465096037 +0000 UTC m=+2054.289831618" Mar 09 09:54:14 crc kubenswrapper[4692]: E0309 09:54:14.212090 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd5610b5_0c3c_4673_9400_b571f4d07531.slice/crio-33dd9b6ec14546fb13067eb98fd438f8a369face0c9c0525fee2717b02c98869.scope\": RecentStats: unable to find data in memory cache]" Mar 09 09:54:14 crc kubenswrapper[4692]: I0309 09:54:14.457902 4692 generic.go:334] "Generic (PLEG): container finished" podID="bd5610b5-0c3c-4673-9400-b571f4d07531" containerID="33dd9b6ec14546fb13067eb98fd438f8a369face0c9c0525fee2717b02c98869" exitCode=0 Mar 09 09:54:14 crc kubenswrapper[4692]: I0309 09:54:14.457952 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" event={"ID":"bd5610b5-0c3c-4673-9400-b571f4d07531","Type":"ContainerDied","Data":"33dd9b6ec14546fb13067eb98fd438f8a369face0c9c0525fee2717b02c98869"} Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.696323 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.726555 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4"] Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.746201 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4"] Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.770651 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-ring-data-devices\") pod \"bd5610b5-0c3c-4673-9400-b571f4d07531\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.770826 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjmlb\" (UniqueName: \"kubernetes.io/projected/bd5610b5-0c3c-4673-9400-b571f4d07531-kube-api-access-bjmlb\") pod \"bd5610b5-0c3c-4673-9400-b571f4d07531\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.771428 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bd5610b5-0c3c-4673-9400-b571f4d07531" (UID: "bd5610b5-0c3c-4673-9400-b571f4d07531"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.771471 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-swiftconf\") pod \"bd5610b5-0c3c-4673-9400-b571f4d07531\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.771616 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-scripts\") pod \"bd5610b5-0c3c-4673-9400-b571f4d07531\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.771659 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-dispersionconf\") pod \"bd5610b5-0c3c-4673-9400-b571f4d07531\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.771757 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd5610b5-0c3c-4673-9400-b571f4d07531-etc-swift\") pod \"bd5610b5-0c3c-4673-9400-b571f4d07531\" (UID: \"bd5610b5-0c3c-4673-9400-b571f4d07531\") " Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.772396 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.773190 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd5610b5-0c3c-4673-9400-b571f4d07531-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bd5610b5-0c3c-4673-9400-b571f4d07531" (UID: "bd5610b5-0c3c-4673-9400-b571f4d07531"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.776188 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd5610b5-0c3c-4673-9400-b571f4d07531-kube-api-access-bjmlb" (OuterVolumeSpecName: "kube-api-access-bjmlb") pod "bd5610b5-0c3c-4673-9400-b571f4d07531" (UID: "bd5610b5-0c3c-4673-9400-b571f4d07531"). InnerVolumeSpecName "kube-api-access-bjmlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.792541 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-scripts" (OuterVolumeSpecName: "scripts") pod "bd5610b5-0c3c-4673-9400-b571f4d07531" (UID: "bd5610b5-0c3c-4673-9400-b571f4d07531"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.795312 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bd5610b5-0c3c-4673-9400-b571f4d07531" (UID: "bd5610b5-0c3c-4673-9400-b571f4d07531"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.795337 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bd5610b5-0c3c-4673-9400-b571f4d07531" (UID: "bd5610b5-0c3c-4673-9400-b571f4d07531"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.874484 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjmlb\" (UniqueName: \"kubernetes.io/projected/bd5610b5-0c3c-4673-9400-b571f4d07531-kube-api-access-bjmlb\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.874528 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.874541 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd5610b5-0c3c-4673-9400-b571f4d07531-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.874553 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bd5610b5-0c3c-4673-9400-b571f4d07531-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:15 crc kubenswrapper[4692]: I0309 09:54:15.874564 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bd5610b5-0c3c-4673-9400-b571f4d07531-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.081337 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd5610b5-0c3c-4673-9400-b571f4d07531" path="/var/lib/kubelet/pods/bd5610b5-0c3c-4673-9400-b571f4d07531/volumes" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.483954 4692 scope.go:117] "RemoveContainer" containerID="33dd9b6ec14546fb13067eb98fd438f8a369face0c9c0525fee2717b02c98869" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.484136 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2c9z4" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.888291 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc"] Mar 09 09:54:16 crc kubenswrapper[4692]: E0309 09:54:16.888776 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd5610b5-0c3c-4673-9400-b571f4d07531" containerName="swift-ring-rebalance" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.888795 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd5610b5-0c3c-4673-9400-b571f4d07531" containerName="swift-ring-rebalance" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.888972 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd5610b5-0c3c-4673-9400-b571f4d07531" containerName="swift-ring-rebalance" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.889969 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.891594 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.892631 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.900317 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc"] Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.990325 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-dispersionconf\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.990408 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2aa40d13-830d-42e7-9381-ba4dbd1e3698-etc-swift\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.990456 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw4r9\" (UniqueName: \"kubernetes.io/projected/2aa40d13-830d-42e7-9381-ba4dbd1e3698-kube-api-access-mw4r9\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.990477 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-ring-data-devices\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.990506 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-scripts\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:16 crc kubenswrapper[4692]: I0309 09:54:16.990530 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-swiftconf\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.091723 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-dispersionconf\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.091802 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2aa40d13-830d-42e7-9381-ba4dbd1e3698-etc-swift\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.091849 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw4r9\" (UniqueName: \"kubernetes.io/projected/2aa40d13-830d-42e7-9381-ba4dbd1e3698-kube-api-access-mw4r9\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.091871 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-ring-data-devices\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.091898 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-scripts\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.091919 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-swiftconf\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.092700 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2aa40d13-830d-42e7-9381-ba4dbd1e3698-etc-swift\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.093299 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-ring-data-devices\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.093399 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-scripts\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.096921 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-swiftconf\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.096922 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-dispersionconf\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.110634 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw4r9\" (UniqueName: \"kubernetes.io/projected/2aa40d13-830d-42e7-9381-ba4dbd1e3698-kube-api-access-mw4r9\") pod \"swift-ring-rebalance-debug-tz4cc\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.209056 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.608220 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc"] Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.623769 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.623817 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.623978 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.625097 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cfb6f9265bcdee5d06e08425c1dd3a025aa1540a263740462908a83a3191a85f"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:54:17 crc kubenswrapper[4692]: I0309 09:54:17.625175 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://cfb6f9265bcdee5d06e08425c1dd3a025aa1540a263740462908a83a3191a85f" gracePeriod=600 Mar 09 09:54:18 crc kubenswrapper[4692]: I0309 09:54:18.501273 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="cfb6f9265bcdee5d06e08425c1dd3a025aa1540a263740462908a83a3191a85f" exitCode=0 Mar 09 09:54:18 crc kubenswrapper[4692]: I0309 09:54:18.501846 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"cfb6f9265bcdee5d06e08425c1dd3a025aa1540a263740462908a83a3191a85f"} Mar 09 09:54:18 crc kubenswrapper[4692]: I0309 09:54:18.501880 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff"} Mar 09 09:54:18 crc kubenswrapper[4692]: I0309 09:54:18.501902 4692 scope.go:117] "RemoveContainer" containerID="14b55ffd155f10ad0a5f3de6586887665abccd582cd4dbf006b3eb4106368c46" Mar 09 09:54:18 crc kubenswrapper[4692]: I0309 09:54:18.504861 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" event={"ID":"2aa40d13-830d-42e7-9381-ba4dbd1e3698","Type":"ContainerStarted","Data":"9100ffdd6a1b366abe731cfd8b84c0588f3568266469b712d02fc54357e3a499"} Mar 09 09:54:18 crc kubenswrapper[4692]: I0309 09:54:18.504903 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" event={"ID":"2aa40d13-830d-42e7-9381-ba4dbd1e3698","Type":"ContainerStarted","Data":"c59266de5ed62465bad90069003b537119dfcf1ffccb1f62c856b3d002544cde"} Mar 09 09:54:18 crc kubenswrapper[4692]: I0309 09:54:18.535778 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" podStartSLOduration=2.535755697 podStartE2EDuration="2.535755697s" podCreationTimestamp="2026-03-09 09:54:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:18.530494228 +0000 UTC m=+2059.355229819" watchObservedRunningTime="2026-03-09 09:54:18.535755697 +0000 UTC m=+2059.360491278" Mar 09 09:54:19 crc kubenswrapper[4692]: I0309 09:54:19.515641 4692 generic.go:334] "Generic (PLEG): container finished" podID="2aa40d13-830d-42e7-9381-ba4dbd1e3698" containerID="9100ffdd6a1b366abe731cfd8b84c0588f3568266469b712d02fc54357e3a499" exitCode=0 Mar 09 09:54:19 crc kubenswrapper[4692]: I0309 09:54:19.515746 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" event={"ID":"2aa40d13-830d-42e7-9381-ba4dbd1e3698","Type":"ContainerDied","Data":"9100ffdd6a1b366abe731cfd8b84c0588f3568266469b712d02fc54357e3a499"} Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.808540 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.838656 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc"] Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.844181 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc"] Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.950328 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-dispersionconf\") pod \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.950390 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2aa40d13-830d-42e7-9381-ba4dbd1e3698-etc-swift\") pod \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.950467 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-swiftconf\") pod \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.950500 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw4r9\" (UniqueName: \"kubernetes.io/projected/2aa40d13-830d-42e7-9381-ba4dbd1e3698-kube-api-access-mw4r9\") pod \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.950529 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-scripts\") pod \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.950574 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-ring-data-devices\") pod \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\" (UID: \"2aa40d13-830d-42e7-9381-ba4dbd1e3698\") " Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.951512 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aa40d13-830d-42e7-9381-ba4dbd1e3698-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2aa40d13-830d-42e7-9381-ba4dbd1e3698" (UID: "2aa40d13-830d-42e7-9381-ba4dbd1e3698"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.951547 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2aa40d13-830d-42e7-9381-ba4dbd1e3698" (UID: "2aa40d13-830d-42e7-9381-ba4dbd1e3698"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.957034 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aa40d13-830d-42e7-9381-ba4dbd1e3698-kube-api-access-mw4r9" (OuterVolumeSpecName: "kube-api-access-mw4r9") pod "2aa40d13-830d-42e7-9381-ba4dbd1e3698" (UID: "2aa40d13-830d-42e7-9381-ba4dbd1e3698"). InnerVolumeSpecName "kube-api-access-mw4r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.971148 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-scripts" (OuterVolumeSpecName: "scripts") pod "2aa40d13-830d-42e7-9381-ba4dbd1e3698" (UID: "2aa40d13-830d-42e7-9381-ba4dbd1e3698"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.975605 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2aa40d13-830d-42e7-9381-ba4dbd1e3698" (UID: "2aa40d13-830d-42e7-9381-ba4dbd1e3698"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:20 crc kubenswrapper[4692]: I0309 09:54:20.976817 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2aa40d13-830d-42e7-9381-ba4dbd1e3698" (UID: "2aa40d13-830d-42e7-9381-ba4dbd1e3698"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.052698 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.052739 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw4r9\" (UniqueName: \"kubernetes.io/projected/2aa40d13-830d-42e7-9381-ba4dbd1e3698-kube-api-access-mw4r9\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.052752 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.052762 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2aa40d13-830d-42e7-9381-ba4dbd1e3698-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.052772 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2aa40d13-830d-42e7-9381-ba4dbd1e3698-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.052784 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2aa40d13-830d-42e7-9381-ba4dbd1e3698-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.537408 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c59266de5ed62465bad90069003b537119dfcf1ffccb1f62c856b3d002544cde" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.537475 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tz4cc" Mar 09 09:54:21 crc kubenswrapper[4692]: I0309 09:54:21.999881 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tznnf"] Mar 09 09:54:22 crc kubenswrapper[4692]: E0309 09:54:22.000270 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aa40d13-830d-42e7-9381-ba4dbd1e3698" containerName="swift-ring-rebalance" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.000357 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aa40d13-830d-42e7-9381-ba4dbd1e3698" containerName="swift-ring-rebalance" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.000537 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aa40d13-830d-42e7-9381-ba4dbd1e3698" containerName="swift-ring-rebalance" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.001177 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.006681 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.006912 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.011955 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tznnf"] Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.067075 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efc2110e-b944-419e-989c-9f23175d61ea-etc-swift\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.067132 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-swiftconf\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.067189 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-scripts\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.067209 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-ring-data-devices\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.067235 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-dispersionconf\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.067270 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrr82\" (UniqueName: \"kubernetes.io/projected/efc2110e-b944-419e-989c-9f23175d61ea-kube-api-access-jrr82\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.080564 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aa40d13-830d-42e7-9381-ba4dbd1e3698" path="/var/lib/kubelet/pods/2aa40d13-830d-42e7-9381-ba4dbd1e3698/volumes" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.168381 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrr82\" (UniqueName: \"kubernetes.io/projected/efc2110e-b944-419e-989c-9f23175d61ea-kube-api-access-jrr82\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.168453 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efc2110e-b944-419e-989c-9f23175d61ea-etc-swift\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.168489 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-swiftconf\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.168525 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-scripts\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.168543 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-ring-data-devices\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.168575 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-dispersionconf\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.169349 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efc2110e-b944-419e-989c-9f23175d61ea-etc-swift\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.169832 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-ring-data-devices\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.170022 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-scripts\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.172798 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-swiftconf\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.176321 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-dispersionconf\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.186972 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrr82\" (UniqueName: \"kubernetes.io/projected/efc2110e-b944-419e-989c-9f23175d61ea-kube-api-access-jrr82\") pod \"swift-ring-rebalance-debug-tznnf\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.375601 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:22 crc kubenswrapper[4692]: I0309 09:54:22.828377 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tznnf"] Mar 09 09:54:22 crc kubenswrapper[4692]: W0309 09:54:22.837436 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefc2110e_b944_419e_989c_9f23175d61ea.slice/crio-38b49c23ecf342869450d869edd856a87de7eff1d083e7e14912bc95c90227b5 WatchSource:0}: Error finding container 38b49c23ecf342869450d869edd856a87de7eff1d083e7e14912bc95c90227b5: Status 404 returned error can't find the container with id 38b49c23ecf342869450d869edd856a87de7eff1d083e7e14912bc95c90227b5 Mar 09 09:54:23 crc kubenswrapper[4692]: I0309 09:54:23.556916 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" event={"ID":"efc2110e-b944-419e-989c-9f23175d61ea","Type":"ContainerStarted","Data":"7045586ad1f6b79f4385d911febbf884f3ecd43f9bd85b376869180547cc750a"} Mar 09 09:54:23 crc kubenswrapper[4692]: I0309 09:54:23.557225 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" event={"ID":"efc2110e-b944-419e-989c-9f23175d61ea","Type":"ContainerStarted","Data":"38b49c23ecf342869450d869edd856a87de7eff1d083e7e14912bc95c90227b5"} Mar 09 09:54:23 crc kubenswrapper[4692]: I0309 09:54:23.584283 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" podStartSLOduration=2.5842611399999997 podStartE2EDuration="2.58426114s" podCreationTimestamp="2026-03-09 09:54:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:23.582187141 +0000 UTC m=+2064.406922722" watchObservedRunningTime="2026-03-09 09:54:23.58426114 +0000 UTC m=+2064.408996721" Mar 09 09:54:24 crc kubenswrapper[4692]: I0309 09:54:24.568065 4692 generic.go:334] "Generic (PLEG): container finished" podID="efc2110e-b944-419e-989c-9f23175d61ea" containerID="7045586ad1f6b79f4385d911febbf884f3ecd43f9bd85b376869180547cc750a" exitCode=0 Mar 09 09:54:24 crc kubenswrapper[4692]: I0309 09:54:24.568220 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" event={"ID":"efc2110e-b944-419e-989c-9f23175d61ea","Type":"ContainerDied","Data":"7045586ad1f6b79f4385d911febbf884f3ecd43f9bd85b376869180547cc750a"} Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.823409 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.871477 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tznnf"] Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.878367 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-tznnf"] Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.921984 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-scripts\") pod \"efc2110e-b944-419e-989c-9f23175d61ea\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.922052 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efc2110e-b944-419e-989c-9f23175d61ea-etc-swift\") pod \"efc2110e-b944-419e-989c-9f23175d61ea\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.922075 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-swiftconf\") pod \"efc2110e-b944-419e-989c-9f23175d61ea\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.922525 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-ring-data-devices\") pod \"efc2110e-b944-419e-989c-9f23175d61ea\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.922613 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrr82\" (UniqueName: \"kubernetes.io/projected/efc2110e-b944-419e-989c-9f23175d61ea-kube-api-access-jrr82\") pod \"efc2110e-b944-419e-989c-9f23175d61ea\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.923025 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "efc2110e-b944-419e-989c-9f23175d61ea" (UID: "efc2110e-b944-419e-989c-9f23175d61ea"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.923180 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efc2110e-b944-419e-989c-9f23175d61ea-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "efc2110e-b944-419e-989c-9f23175d61ea" (UID: "efc2110e-b944-419e-989c-9f23175d61ea"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.923240 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-dispersionconf\") pod \"efc2110e-b944-419e-989c-9f23175d61ea\" (UID: \"efc2110e-b944-419e-989c-9f23175d61ea\") " Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.923575 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/efc2110e-b944-419e-989c-9f23175d61ea-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.923590 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.927956 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efc2110e-b944-419e-989c-9f23175d61ea-kube-api-access-jrr82" (OuterVolumeSpecName: "kube-api-access-jrr82") pod "efc2110e-b944-419e-989c-9f23175d61ea" (UID: "efc2110e-b944-419e-989c-9f23175d61ea"). InnerVolumeSpecName "kube-api-access-jrr82". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.944445 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-scripts" (OuterVolumeSpecName: "scripts") pod "efc2110e-b944-419e-989c-9f23175d61ea" (UID: "efc2110e-b944-419e-989c-9f23175d61ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.945898 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "efc2110e-b944-419e-989c-9f23175d61ea" (UID: "efc2110e-b944-419e-989c-9f23175d61ea"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:25 crc kubenswrapper[4692]: I0309 09:54:25.950761 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "efc2110e-b944-419e-989c-9f23175d61ea" (UID: "efc2110e-b944-419e-989c-9f23175d61ea"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:26 crc kubenswrapper[4692]: I0309 09:54:26.025404 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efc2110e-b944-419e-989c-9f23175d61ea-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:26 crc kubenswrapper[4692]: I0309 09:54:26.025452 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:26 crc kubenswrapper[4692]: I0309 09:54:26.025463 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrr82\" (UniqueName: \"kubernetes.io/projected/efc2110e-b944-419e-989c-9f23175d61ea-kube-api-access-jrr82\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:26 crc kubenswrapper[4692]: I0309 09:54:26.025475 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/efc2110e-b944-419e-989c-9f23175d61ea-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:26 crc kubenswrapper[4692]: I0309 09:54:26.080431 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efc2110e-b944-419e-989c-9f23175d61ea" path="/var/lib/kubelet/pods/efc2110e-b944-419e-989c-9f23175d61ea/volumes" Mar 09 09:54:26 crc kubenswrapper[4692]: I0309 09:54:26.584767 4692 scope.go:117] "RemoveContainer" containerID="7045586ad1f6b79f4385d911febbf884f3ecd43f9bd85b376869180547cc750a" Mar 09 09:54:26 crc kubenswrapper[4692]: I0309 09:54:26.584903 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-tznnf" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.040090 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q44jw"] Mar 09 09:54:27 crc kubenswrapper[4692]: E0309 09:54:27.040566 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efc2110e-b944-419e-989c-9f23175d61ea" containerName="swift-ring-rebalance" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.040596 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="efc2110e-b944-419e-989c-9f23175d61ea" containerName="swift-ring-rebalance" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.040836 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="efc2110e-b944-419e-989c-9f23175d61ea" containerName="swift-ring-rebalance" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.041492 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.044220 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.050257 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q44jw"] Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.050276 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.142070 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/faad4007-6043-416c-8c7a-cce7e6738f65-etc-swift\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.142148 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-ring-data-devices\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.142205 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct6vl\" (UniqueName: \"kubernetes.io/projected/faad4007-6043-416c-8c7a-cce7e6738f65-kube-api-access-ct6vl\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.142403 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-swiftconf\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.142663 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-scripts\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.142879 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-dispersionconf\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.244576 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/faad4007-6043-416c-8c7a-cce7e6738f65-etc-swift\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.244636 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-ring-data-devices\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.244676 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct6vl\" (UniqueName: \"kubernetes.io/projected/faad4007-6043-416c-8c7a-cce7e6738f65-kube-api-access-ct6vl\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.245124 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-swiftconf\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.245207 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/faad4007-6043-416c-8c7a-cce7e6738f65-etc-swift\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.245247 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-scripts\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.245342 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-dispersionconf\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.245653 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-ring-data-devices\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.245784 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-scripts\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.251342 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-swiftconf\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.256708 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-dispersionconf\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.261218 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct6vl\" (UniqueName: \"kubernetes.io/projected/faad4007-6043-416c-8c7a-cce7e6738f65-kube-api-access-ct6vl\") pod \"swift-ring-rebalance-debug-q44jw\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.358014 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:27 crc kubenswrapper[4692]: I0309 09:54:27.783419 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q44jw"] Mar 09 09:54:28 crc kubenswrapper[4692]: I0309 09:54:28.601709 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" event={"ID":"faad4007-6043-416c-8c7a-cce7e6738f65","Type":"ContainerStarted","Data":"e99aa3985d9f9af6c359f6a6b82df125a0064c276634fa986fb9014f44f47afa"} Mar 09 09:54:28 crc kubenswrapper[4692]: I0309 09:54:28.602053 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" event={"ID":"faad4007-6043-416c-8c7a-cce7e6738f65","Type":"ContainerStarted","Data":"77196dce1f8eb49bf776cf8d080fc04b38be9cbfb01741441738b477bc638d0d"} Mar 09 09:54:28 crc kubenswrapper[4692]: I0309 09:54:28.630005 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" podStartSLOduration=1.6299846059999998 podStartE2EDuration="1.629984606s" podCreationTimestamp="2026-03-09 09:54:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:28.628683359 +0000 UTC m=+2069.453418970" watchObservedRunningTime="2026-03-09 09:54:28.629984606 +0000 UTC m=+2069.454720197" Mar 09 09:54:29 crc kubenswrapper[4692]: I0309 09:54:29.611857 4692 generic.go:334] "Generic (PLEG): container finished" podID="faad4007-6043-416c-8c7a-cce7e6738f65" containerID="e99aa3985d9f9af6c359f6a6b82df125a0064c276634fa986fb9014f44f47afa" exitCode=0 Mar 09 09:54:29 crc kubenswrapper[4692]: I0309 09:54:29.611916 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" event={"ID":"faad4007-6043-416c-8c7a-cce7e6738f65","Type":"ContainerDied","Data":"e99aa3985d9f9af6c359f6a6b82df125a0064c276634fa986fb9014f44f47afa"} Mar 09 09:54:30 crc kubenswrapper[4692]: I0309 09:54:30.871680 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:30 crc kubenswrapper[4692]: I0309 09:54:30.904208 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q44jw"] Mar 09 09:54:30 crc kubenswrapper[4692]: I0309 09:54:30.911543 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-q44jw"] Mar 09 09:54:30 crc kubenswrapper[4692]: I0309 09:54:30.999646 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct6vl\" (UniqueName: \"kubernetes.io/projected/faad4007-6043-416c-8c7a-cce7e6738f65-kube-api-access-ct6vl\") pod \"faad4007-6043-416c-8c7a-cce7e6738f65\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " Mar 09 09:54:30 crc kubenswrapper[4692]: I0309 09:54:30.999699 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-swiftconf\") pod \"faad4007-6043-416c-8c7a-cce7e6738f65\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:30.999842 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/faad4007-6043-416c-8c7a-cce7e6738f65-etc-swift\") pod \"faad4007-6043-416c-8c7a-cce7e6738f65\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:30.999868 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-ring-data-devices\") pod \"faad4007-6043-416c-8c7a-cce7e6738f65\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:30.999894 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-dispersionconf\") pod \"faad4007-6043-416c-8c7a-cce7e6738f65\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:30.999936 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-scripts\") pod \"faad4007-6043-416c-8c7a-cce7e6738f65\" (UID: \"faad4007-6043-416c-8c7a-cce7e6738f65\") " Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.000799 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "faad4007-6043-416c-8c7a-cce7e6738f65" (UID: "faad4007-6043-416c-8c7a-cce7e6738f65"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.001112 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faad4007-6043-416c-8c7a-cce7e6738f65-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "faad4007-6043-416c-8c7a-cce7e6738f65" (UID: "faad4007-6043-416c-8c7a-cce7e6738f65"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.005202 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faad4007-6043-416c-8c7a-cce7e6738f65-kube-api-access-ct6vl" (OuterVolumeSpecName: "kube-api-access-ct6vl") pod "faad4007-6043-416c-8c7a-cce7e6738f65" (UID: "faad4007-6043-416c-8c7a-cce7e6738f65"). InnerVolumeSpecName "kube-api-access-ct6vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.022780 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-scripts" (OuterVolumeSpecName: "scripts") pod "faad4007-6043-416c-8c7a-cce7e6738f65" (UID: "faad4007-6043-416c-8c7a-cce7e6738f65"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.024285 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "faad4007-6043-416c-8c7a-cce7e6738f65" (UID: "faad4007-6043-416c-8c7a-cce7e6738f65"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.024539 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "faad4007-6043-416c-8c7a-cce7e6738f65" (UID: "faad4007-6043-416c-8c7a-cce7e6738f65"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.101824 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/faad4007-6043-416c-8c7a-cce7e6738f65-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.101856 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.101866 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.101875 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/faad4007-6043-416c-8c7a-cce7e6738f65-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.101884 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct6vl\" (UniqueName: \"kubernetes.io/projected/faad4007-6043-416c-8c7a-cce7e6738f65-kube-api-access-ct6vl\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.101893 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/faad4007-6043-416c-8c7a-cce7e6738f65-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.628846 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77196dce1f8eb49bf776cf8d080fc04b38be9cbfb01741441738b477bc638d0d" Mar 09 09:54:31 crc kubenswrapper[4692]: I0309 09:54:31.628902 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-q44jw" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.055737 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl"] Mar 09 09:54:32 crc kubenswrapper[4692]: E0309 09:54:32.056015 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faad4007-6043-416c-8c7a-cce7e6738f65" containerName="swift-ring-rebalance" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.056027 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="faad4007-6043-416c-8c7a-cce7e6738f65" containerName="swift-ring-rebalance" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.056251 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="faad4007-6043-416c-8c7a-cce7e6738f65" containerName="swift-ring-rebalance" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.056749 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.058968 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.059041 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.068517 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl"] Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.095007 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faad4007-6043-416c-8c7a-cce7e6738f65" path="/var/lib/kubelet/pods/faad4007-6043-416c-8c7a-cce7e6738f65/volumes" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.116995 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-scripts\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.117093 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-ring-data-devices\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.117139 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-dispersionconf\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.117179 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-swiftconf\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.117227 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z54wl\" (UniqueName: \"kubernetes.io/projected/47d10837-9bba-4276-9aa5-67e3e6836219-kube-api-access-z54wl\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.117253 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47d10837-9bba-4276-9aa5-67e3e6836219-etc-swift\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.219633 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-ring-data-devices\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.219713 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-dispersionconf\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.219730 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-swiftconf\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.219763 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z54wl\" (UniqueName: \"kubernetes.io/projected/47d10837-9bba-4276-9aa5-67e3e6836219-kube-api-access-z54wl\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.219797 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47d10837-9bba-4276-9aa5-67e3e6836219-etc-swift\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.219851 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-scripts\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.220652 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-scripts\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.220649 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-ring-data-devices\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.221296 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47d10837-9bba-4276-9aa5-67e3e6836219-etc-swift\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.225060 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-dispersionconf\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.225185 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-swiftconf\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.238735 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z54wl\" (UniqueName: \"kubernetes.io/projected/47d10837-9bba-4276-9aa5-67e3e6836219-kube-api-access-z54wl\") pod \"swift-ring-rebalance-debug-sxgbl\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.389782 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:32 crc kubenswrapper[4692]: I0309 09:54:32.824976 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl"] Mar 09 09:54:32 crc kubenswrapper[4692]: W0309 09:54:32.828181 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47d10837_9bba_4276_9aa5_67e3e6836219.slice/crio-c15e41fa0583a0779672740d7af37916ae581b7c9175b0da6b97bcd3bf25a279 WatchSource:0}: Error finding container c15e41fa0583a0779672740d7af37916ae581b7c9175b0da6b97bcd3bf25a279: Status 404 returned error can't find the container with id c15e41fa0583a0779672740d7af37916ae581b7c9175b0da6b97bcd3bf25a279 Mar 09 09:54:33 crc kubenswrapper[4692]: I0309 09:54:33.648774 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" event={"ID":"47d10837-9bba-4276-9aa5-67e3e6836219","Type":"ContainerStarted","Data":"9fdc08c4cb6f33d8c1babf320f4e37c65adf69ad504ceb6a71dd06a9ba5316df"} Mar 09 09:54:33 crc kubenswrapper[4692]: I0309 09:54:33.649111 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" event={"ID":"47d10837-9bba-4276-9aa5-67e3e6836219","Type":"ContainerStarted","Data":"c15e41fa0583a0779672740d7af37916ae581b7c9175b0da6b97bcd3bf25a279"} Mar 09 09:54:34 crc kubenswrapper[4692]: I0309 09:54:34.658124 4692 generic.go:334] "Generic (PLEG): container finished" podID="47d10837-9bba-4276-9aa5-67e3e6836219" containerID="9fdc08c4cb6f33d8c1babf320f4e37c65adf69ad504ceb6a71dd06a9ba5316df" exitCode=0 Mar 09 09:54:34 crc kubenswrapper[4692]: I0309 09:54:34.658292 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" event={"ID":"47d10837-9bba-4276-9aa5-67e3e6836219","Type":"ContainerDied","Data":"9fdc08c4cb6f33d8c1babf320f4e37c65adf69ad504ceb6a71dd06a9ba5316df"} Mar 09 09:54:35 crc kubenswrapper[4692]: I0309 09:54:35.957926 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.009325 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl"] Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.015090 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl"] Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.075524 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-dispersionconf\") pod \"47d10837-9bba-4276-9aa5-67e3e6836219\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.075612 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47d10837-9bba-4276-9aa5-67e3e6836219-etc-swift\") pod \"47d10837-9bba-4276-9aa5-67e3e6836219\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.075640 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-swiftconf\") pod \"47d10837-9bba-4276-9aa5-67e3e6836219\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.075673 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-ring-data-devices\") pod \"47d10837-9bba-4276-9aa5-67e3e6836219\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.075742 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z54wl\" (UniqueName: \"kubernetes.io/projected/47d10837-9bba-4276-9aa5-67e3e6836219-kube-api-access-z54wl\") pod \"47d10837-9bba-4276-9aa5-67e3e6836219\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.075765 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-scripts\") pod \"47d10837-9bba-4276-9aa5-67e3e6836219\" (UID: \"47d10837-9bba-4276-9aa5-67e3e6836219\") " Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.077136 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "47d10837-9bba-4276-9aa5-67e3e6836219" (UID: "47d10837-9bba-4276-9aa5-67e3e6836219"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.077213 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47d10837-9bba-4276-9aa5-67e3e6836219-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "47d10837-9bba-4276-9aa5-67e3e6836219" (UID: "47d10837-9bba-4276-9aa5-67e3e6836219"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.081345 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d10837-9bba-4276-9aa5-67e3e6836219-kube-api-access-z54wl" (OuterVolumeSpecName: "kube-api-access-z54wl") pod "47d10837-9bba-4276-9aa5-67e3e6836219" (UID: "47d10837-9bba-4276-9aa5-67e3e6836219"). InnerVolumeSpecName "kube-api-access-z54wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.096707 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-scripts" (OuterVolumeSpecName: "scripts") pod "47d10837-9bba-4276-9aa5-67e3e6836219" (UID: "47d10837-9bba-4276-9aa5-67e3e6836219"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.098742 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "47d10837-9bba-4276-9aa5-67e3e6836219" (UID: "47d10837-9bba-4276-9aa5-67e3e6836219"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.103478 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "47d10837-9bba-4276-9aa5-67e3e6836219" (UID: "47d10837-9bba-4276-9aa5-67e3e6836219"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.177182 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.177220 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z54wl\" (UniqueName: \"kubernetes.io/projected/47d10837-9bba-4276-9aa5-67e3e6836219-kube-api-access-z54wl\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.177237 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47d10837-9bba-4276-9aa5-67e3e6836219-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.177248 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.177258 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47d10837-9bba-4276-9aa5-67e3e6836219-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.177268 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47d10837-9bba-4276-9aa5-67e3e6836219-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.674456 4692 scope.go:117] "RemoveContainer" containerID="9fdc08c4cb6f33d8c1babf320f4e37c65adf69ad504ceb6a71dd06a9ba5316df" Mar 09 09:54:36 crc kubenswrapper[4692]: I0309 09:54:36.674507 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sxgbl" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.148190 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9"] Mar 09 09:54:37 crc kubenswrapper[4692]: E0309 09:54:37.149664 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d10837-9bba-4276-9aa5-67e3e6836219" containerName="swift-ring-rebalance" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.149763 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d10837-9bba-4276-9aa5-67e3e6836219" containerName="swift-ring-rebalance" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.150026 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d10837-9bba-4276-9aa5-67e3e6836219" containerName="swift-ring-rebalance" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.150780 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.153531 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.153755 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.156223 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9"] Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.199106 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-scripts\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.199198 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.199257 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-dispersionconf\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.199282 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-swiftconf\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.199310 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csh9j\" (UniqueName: \"kubernetes.io/projected/c937b505-397f-4be0-b8f1-de536f05e930-kube-api-access-csh9j\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.199330 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c937b505-397f-4be0-b8f1-de536f05e930-etc-swift\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.301120 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csh9j\" (UniqueName: \"kubernetes.io/projected/c937b505-397f-4be0-b8f1-de536f05e930-kube-api-access-csh9j\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.301175 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c937b505-397f-4be0-b8f1-de536f05e930-etc-swift\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.301234 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-scripts\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.301265 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.301304 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-dispersionconf\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.301324 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-swiftconf\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.301924 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c937b505-397f-4be0-b8f1-de536f05e930-etc-swift\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.302155 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-scripts\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.302359 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.305454 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-dispersionconf\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.305454 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-swiftconf\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.315926 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csh9j\" (UniqueName: \"kubernetes.io/projected/c937b505-397f-4be0-b8f1-de536f05e930-kube-api-access-csh9j\") pod \"swift-ring-rebalance-debug-j4pz9\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.473835 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:37 crc kubenswrapper[4692]: I0309 09:54:37.783056 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9"] Mar 09 09:54:38 crc kubenswrapper[4692]: I0309 09:54:38.081062 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47d10837-9bba-4276-9aa5-67e3e6836219" path="/var/lib/kubelet/pods/47d10837-9bba-4276-9aa5-67e3e6836219/volumes" Mar 09 09:54:38 crc kubenswrapper[4692]: I0309 09:54:38.730553 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" event={"ID":"c937b505-397f-4be0-b8f1-de536f05e930","Type":"ContainerStarted","Data":"a0cd5a083aca529493f9715f1fc4bd83fd47ccbc981f17d396663bdabf8682f2"} Mar 09 09:54:38 crc kubenswrapper[4692]: I0309 09:54:38.730857 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" event={"ID":"c937b505-397f-4be0-b8f1-de536f05e930","Type":"ContainerStarted","Data":"e8f163ce2325aba25aa1f732b2da8eebdca2fbb5524235d3c1a1af0c5ce281ad"} Mar 09 09:54:38 crc kubenswrapper[4692]: I0309 09:54:38.751558 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" podStartSLOduration=1.751541655 podStartE2EDuration="1.751541655s" podCreationTimestamp="2026-03-09 09:54:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:38.750801504 +0000 UTC m=+2079.575537095" watchObservedRunningTime="2026-03-09 09:54:38.751541655 +0000 UTC m=+2079.576277246" Mar 09 09:54:39 crc kubenswrapper[4692]: I0309 09:54:39.742258 4692 generic.go:334] "Generic (PLEG): container finished" podID="c937b505-397f-4be0-b8f1-de536f05e930" containerID="a0cd5a083aca529493f9715f1fc4bd83fd47ccbc981f17d396663bdabf8682f2" exitCode=0 Mar 09 09:54:39 crc kubenswrapper[4692]: I0309 09:54:39.742309 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" event={"ID":"c937b505-397f-4be0-b8f1-de536f05e930","Type":"ContainerDied","Data":"a0cd5a083aca529493f9715f1fc4bd83fd47ccbc981f17d396663bdabf8682f2"} Mar 09 09:54:40 crc kubenswrapper[4692]: I0309 09:54:40.980613 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.016979 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9"] Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.022083 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9"] Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.055506 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-dispersionconf\") pod \"c937b505-397f-4be0-b8f1-de536f05e930\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.055593 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-scripts\") pod \"c937b505-397f-4be0-b8f1-de536f05e930\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.055648 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-swiftconf\") pod \"c937b505-397f-4be0-b8f1-de536f05e930\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.055678 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c937b505-397f-4be0-b8f1-de536f05e930-etc-swift\") pod \"c937b505-397f-4be0-b8f1-de536f05e930\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.055723 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-ring-data-devices\") pod \"c937b505-397f-4be0-b8f1-de536f05e930\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.055757 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csh9j\" (UniqueName: \"kubernetes.io/projected/c937b505-397f-4be0-b8f1-de536f05e930-kube-api-access-csh9j\") pod \"c937b505-397f-4be0-b8f1-de536f05e930\" (UID: \"c937b505-397f-4be0-b8f1-de536f05e930\") " Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.056649 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c937b505-397f-4be0-b8f1-de536f05e930" (UID: "c937b505-397f-4be0-b8f1-de536f05e930"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.056942 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c937b505-397f-4be0-b8f1-de536f05e930-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c937b505-397f-4be0-b8f1-de536f05e930" (UID: "c937b505-397f-4be0-b8f1-de536f05e930"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.061448 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c937b505-397f-4be0-b8f1-de536f05e930-kube-api-access-csh9j" (OuterVolumeSpecName: "kube-api-access-csh9j") pod "c937b505-397f-4be0-b8f1-de536f05e930" (UID: "c937b505-397f-4be0-b8f1-de536f05e930"). InnerVolumeSpecName "kube-api-access-csh9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.076342 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c937b505-397f-4be0-b8f1-de536f05e930" (UID: "c937b505-397f-4be0-b8f1-de536f05e930"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.077678 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-scripts" (OuterVolumeSpecName: "scripts") pod "c937b505-397f-4be0-b8f1-de536f05e930" (UID: "c937b505-397f-4be0-b8f1-de536f05e930"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.094202 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c937b505-397f-4be0-b8f1-de536f05e930" (UID: "c937b505-397f-4be0-b8f1-de536f05e930"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.157059 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.157098 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.157237 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c937b505-397f-4be0-b8f1-de536f05e930-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.157253 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c937b505-397f-4be0-b8f1-de536f05e930-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.157269 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csh9j\" (UniqueName: \"kubernetes.io/projected/c937b505-397f-4be0-b8f1-de536f05e930-kube-api-access-csh9j\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.157280 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c937b505-397f-4be0-b8f1-de536f05e930-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.758179 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8f163ce2325aba25aa1f732b2da8eebdca2fbb5524235d3c1a1af0c5ce281ad" Mar 09 09:54:41 crc kubenswrapper[4692]: I0309 09:54:41.758253 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4pz9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.080929 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c937b505-397f-4be0-b8f1-de536f05e930" path="/var/lib/kubelet/pods/c937b505-397f-4be0-b8f1-de536f05e930/volumes" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.154013 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9"] Mar 09 09:54:42 crc kubenswrapper[4692]: E0309 09:54:42.154385 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c937b505-397f-4be0-b8f1-de536f05e930" containerName="swift-ring-rebalance" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.154398 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c937b505-397f-4be0-b8f1-de536f05e930" containerName="swift-ring-rebalance" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.154595 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c937b505-397f-4be0-b8f1-de536f05e930" containerName="swift-ring-rebalance" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.155215 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.156944 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.158618 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.163204 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9"] Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.275098 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-dispersionconf\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.275196 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-ring-data-devices\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.275226 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmpl9\" (UniqueName: \"kubernetes.io/projected/2d1316cc-a757-4afd-83e5-d12ea1637602-kube-api-access-kmpl9\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.275257 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d1316cc-a757-4afd-83e5-d12ea1637602-etc-swift\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.275288 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-swiftconf\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.275359 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-scripts\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.377182 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-dispersionconf\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.377232 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-ring-data-devices\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.377257 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmpl9\" (UniqueName: \"kubernetes.io/projected/2d1316cc-a757-4afd-83e5-d12ea1637602-kube-api-access-kmpl9\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.377283 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d1316cc-a757-4afd-83e5-d12ea1637602-etc-swift\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.377304 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-swiftconf\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.377326 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-scripts\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.378049 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d1316cc-a757-4afd-83e5-d12ea1637602-etc-swift\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.378222 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-scripts\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.378268 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-ring-data-devices\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.381090 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-swiftconf\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.381225 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-dispersionconf\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.393602 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmpl9\" (UniqueName: \"kubernetes.io/projected/2d1316cc-a757-4afd-83e5-d12ea1637602-kube-api-access-kmpl9\") pod \"swift-ring-rebalance-debug-gqzl9\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.478570 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:42 crc kubenswrapper[4692]: I0309 09:54:42.940932 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9"] Mar 09 09:54:43 crc kubenswrapper[4692]: I0309 09:54:43.776158 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" event={"ID":"2d1316cc-a757-4afd-83e5-d12ea1637602","Type":"ContainerStarted","Data":"2bf1f642be46a1f3d093f0f8c44dcae82602cde0cc59a31ee9b2330b1a22f29b"} Mar 09 09:54:43 crc kubenswrapper[4692]: I0309 09:54:43.776581 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" event={"ID":"2d1316cc-a757-4afd-83e5-d12ea1637602","Type":"ContainerStarted","Data":"392f50c7cf3cd5b07464c085d20233a2bd9ef1a2de52cbaebc2996daedb88d4f"} Mar 09 09:54:44 crc kubenswrapper[4692]: I0309 09:54:44.787201 4692 generic.go:334] "Generic (PLEG): container finished" podID="2d1316cc-a757-4afd-83e5-d12ea1637602" containerID="2bf1f642be46a1f3d093f0f8c44dcae82602cde0cc59a31ee9b2330b1a22f29b" exitCode=0 Mar 09 09:54:44 crc kubenswrapper[4692]: I0309 09:54:44.787262 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" event={"ID":"2d1316cc-a757-4afd-83e5-d12ea1637602","Type":"ContainerDied","Data":"2bf1f642be46a1f3d093f0f8c44dcae82602cde0cc59a31ee9b2330b1a22f29b"} Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.081290 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.111985 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9"] Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.118057 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9"] Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.176653 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmpl9\" (UniqueName: \"kubernetes.io/projected/2d1316cc-a757-4afd-83e5-d12ea1637602-kube-api-access-kmpl9\") pod \"2d1316cc-a757-4afd-83e5-d12ea1637602\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.176796 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-scripts\") pod \"2d1316cc-a757-4afd-83e5-d12ea1637602\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.176825 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-swiftconf\") pod \"2d1316cc-a757-4afd-83e5-d12ea1637602\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.176866 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-ring-data-devices\") pod \"2d1316cc-a757-4afd-83e5-d12ea1637602\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.176921 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d1316cc-a757-4afd-83e5-d12ea1637602-etc-swift\") pod \"2d1316cc-a757-4afd-83e5-d12ea1637602\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.176947 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-dispersionconf\") pod \"2d1316cc-a757-4afd-83e5-d12ea1637602\" (UID: \"2d1316cc-a757-4afd-83e5-d12ea1637602\") " Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.177622 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d1316cc-a757-4afd-83e5-d12ea1637602-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2d1316cc-a757-4afd-83e5-d12ea1637602" (UID: "2d1316cc-a757-4afd-83e5-d12ea1637602"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.178438 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2d1316cc-a757-4afd-83e5-d12ea1637602" (UID: "2d1316cc-a757-4afd-83e5-d12ea1637602"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.185789 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d1316cc-a757-4afd-83e5-d12ea1637602-kube-api-access-kmpl9" (OuterVolumeSpecName: "kube-api-access-kmpl9") pod "2d1316cc-a757-4afd-83e5-d12ea1637602" (UID: "2d1316cc-a757-4afd-83e5-d12ea1637602"). InnerVolumeSpecName "kube-api-access-kmpl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.198922 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-scripts" (OuterVolumeSpecName: "scripts") pod "2d1316cc-a757-4afd-83e5-d12ea1637602" (UID: "2d1316cc-a757-4afd-83e5-d12ea1637602"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.218416 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2d1316cc-a757-4afd-83e5-d12ea1637602" (UID: "2d1316cc-a757-4afd-83e5-d12ea1637602"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.234386 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2d1316cc-a757-4afd-83e5-d12ea1637602" (UID: "2d1316cc-a757-4afd-83e5-d12ea1637602"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.279131 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.279200 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.279215 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2d1316cc-a757-4afd-83e5-d12ea1637602-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.279225 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2d1316cc-a757-4afd-83e5-d12ea1637602-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.279234 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2d1316cc-a757-4afd-83e5-d12ea1637602-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.279243 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmpl9\" (UniqueName: \"kubernetes.io/projected/2d1316cc-a757-4afd-83e5-d12ea1637602-kube-api-access-kmpl9\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.802490 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="392f50c7cf3cd5b07464c085d20233a2bd9ef1a2de52cbaebc2996daedb88d4f" Mar 09 09:54:46 crc kubenswrapper[4692]: I0309 09:54:46.802549 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gqzl9" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.265853 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pzswt"] Mar 09 09:54:47 crc kubenswrapper[4692]: E0309 09:54:47.266153 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1316cc-a757-4afd-83e5-d12ea1637602" containerName="swift-ring-rebalance" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.266182 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1316cc-a757-4afd-83e5-d12ea1637602" containerName="swift-ring-rebalance" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.266315 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1316cc-a757-4afd-83e5-d12ea1637602" containerName="swift-ring-rebalance" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.266830 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.268747 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.273833 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.274528 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pzswt"] Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.394772 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-ring-data-devices\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.394919 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-scripts\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.394957 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-dispersionconf\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.394986 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtz2v\" (UniqueName: \"kubernetes.io/projected/4c39d0e5-e889-41b0-9416-984e96a29022-kube-api-access-wtz2v\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.395019 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c39d0e5-e889-41b0-9416-984e96a29022-etc-swift\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.395053 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-swiftconf\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.496091 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-scripts\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.496140 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-dispersionconf\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.496177 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtz2v\" (UniqueName: \"kubernetes.io/projected/4c39d0e5-e889-41b0-9416-984e96a29022-kube-api-access-wtz2v\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.496205 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c39d0e5-e889-41b0-9416-984e96a29022-etc-swift\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.496227 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-swiftconf\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.496260 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-ring-data-devices\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.497035 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c39d0e5-e889-41b0-9416-984e96a29022-etc-swift\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.497078 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-scripts\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.497139 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-ring-data-devices\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.502205 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-dispersionconf\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.502625 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-swiftconf\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.512329 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtz2v\" (UniqueName: \"kubernetes.io/projected/4c39d0e5-e889-41b0-9416-984e96a29022-kube-api-access-wtz2v\") pod \"swift-ring-rebalance-debug-pzswt\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.590072 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:47 crc kubenswrapper[4692]: I0309 09:54:47.983066 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pzswt"] Mar 09 09:54:47 crc kubenswrapper[4692]: W0309 09:54:47.999338 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c39d0e5_e889_41b0_9416_984e96a29022.slice/crio-eefbade4bd2a521277bac78e2b12d549766887c6ad2a81a9c0bab11ccd50413a WatchSource:0}: Error finding container eefbade4bd2a521277bac78e2b12d549766887c6ad2a81a9c0bab11ccd50413a: Status 404 returned error can't find the container with id eefbade4bd2a521277bac78e2b12d549766887c6ad2a81a9c0bab11ccd50413a Mar 09 09:54:48 crc kubenswrapper[4692]: I0309 09:54:48.084003 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d1316cc-a757-4afd-83e5-d12ea1637602" path="/var/lib/kubelet/pods/2d1316cc-a757-4afd-83e5-d12ea1637602/volumes" Mar 09 09:54:48 crc kubenswrapper[4692]: I0309 09:54:48.821535 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" event={"ID":"4c39d0e5-e889-41b0-9416-984e96a29022","Type":"ContainerStarted","Data":"b7875fa5116cc38b75a9fd03aa64eab7c165c848b2e7a772b63e69b7be9e95c8"} Mar 09 09:54:48 crc kubenswrapper[4692]: I0309 09:54:48.821576 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" event={"ID":"4c39d0e5-e889-41b0-9416-984e96a29022","Type":"ContainerStarted","Data":"eefbade4bd2a521277bac78e2b12d549766887c6ad2a81a9c0bab11ccd50413a"} Mar 09 09:54:48 crc kubenswrapper[4692]: I0309 09:54:48.842632 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" podStartSLOduration=1.842609953 podStartE2EDuration="1.842609953s" podCreationTimestamp="2026-03-09 09:54:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:48.838987531 +0000 UTC m=+2089.663723132" watchObservedRunningTime="2026-03-09 09:54:48.842609953 +0000 UTC m=+2089.667345544" Mar 09 09:54:49 crc kubenswrapper[4692]: I0309 09:54:49.830087 4692 generic.go:334] "Generic (PLEG): container finished" podID="4c39d0e5-e889-41b0-9416-984e96a29022" containerID="b7875fa5116cc38b75a9fd03aa64eab7c165c848b2e7a772b63e69b7be9e95c8" exitCode=0 Mar 09 09:54:49 crc kubenswrapper[4692]: I0309 09:54:49.830212 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" event={"ID":"4c39d0e5-e889-41b0-9416-984e96a29022","Type":"ContainerDied","Data":"b7875fa5116cc38b75a9fd03aa64eab7c165c848b2e7a772b63e69b7be9e95c8"} Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.116014 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.154582 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pzswt"] Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.161048 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pzswt"] Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.251711 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-swiftconf\") pod \"4c39d0e5-e889-41b0-9416-984e96a29022\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.251790 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtz2v\" (UniqueName: \"kubernetes.io/projected/4c39d0e5-e889-41b0-9416-984e96a29022-kube-api-access-wtz2v\") pod \"4c39d0e5-e889-41b0-9416-984e96a29022\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.251813 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-dispersionconf\") pod \"4c39d0e5-e889-41b0-9416-984e96a29022\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.251845 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c39d0e5-e889-41b0-9416-984e96a29022-etc-swift\") pod \"4c39d0e5-e889-41b0-9416-984e96a29022\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.251879 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-scripts\") pod \"4c39d0e5-e889-41b0-9416-984e96a29022\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.251964 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-ring-data-devices\") pod \"4c39d0e5-e889-41b0-9416-984e96a29022\" (UID: \"4c39d0e5-e889-41b0-9416-984e96a29022\") " Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.253463 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4c39d0e5-e889-41b0-9416-984e96a29022" (UID: "4c39d0e5-e889-41b0-9416-984e96a29022"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.253495 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c39d0e5-e889-41b0-9416-984e96a29022-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4c39d0e5-e889-41b0-9416-984e96a29022" (UID: "4c39d0e5-e889-41b0-9416-984e96a29022"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.258290 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c39d0e5-e889-41b0-9416-984e96a29022-kube-api-access-wtz2v" (OuterVolumeSpecName: "kube-api-access-wtz2v") pod "4c39d0e5-e889-41b0-9416-984e96a29022" (UID: "4c39d0e5-e889-41b0-9416-984e96a29022"). InnerVolumeSpecName "kube-api-access-wtz2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.280079 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-scripts" (OuterVolumeSpecName: "scripts") pod "4c39d0e5-e889-41b0-9416-984e96a29022" (UID: "4c39d0e5-e889-41b0-9416-984e96a29022"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.280613 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4c39d0e5-e889-41b0-9416-984e96a29022" (UID: "4c39d0e5-e889-41b0-9416-984e96a29022"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.289411 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4c39d0e5-e889-41b0-9416-984e96a29022" (UID: "4c39d0e5-e889-41b0-9416-984e96a29022"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.353414 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.353452 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.353464 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtz2v\" (UniqueName: \"kubernetes.io/projected/4c39d0e5-e889-41b0-9416-984e96a29022-kube-api-access-wtz2v\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.353474 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4c39d0e5-e889-41b0-9416-984e96a29022-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.353482 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4c39d0e5-e889-41b0-9416-984e96a29022-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.353493 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c39d0e5-e889-41b0-9416-984e96a29022-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.849636 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eefbade4bd2a521277bac78e2b12d549766887c6ad2a81a9c0bab11ccd50413a" Mar 09 09:54:51 crc kubenswrapper[4692]: I0309 09:54:51.849693 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pzswt" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.080651 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c39d0e5-e889-41b0-9416-984e96a29022" path="/var/lib/kubelet/pods/4c39d0e5-e889-41b0-9416-984e96a29022/volumes" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.304918 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj"] Mar 09 09:54:52 crc kubenswrapper[4692]: E0309 09:54:52.313606 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c39d0e5-e889-41b0-9416-984e96a29022" containerName="swift-ring-rebalance" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.313642 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c39d0e5-e889-41b0-9416-984e96a29022" containerName="swift-ring-rebalance" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.313896 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c39d0e5-e889-41b0-9416-984e96a29022" containerName="swift-ring-rebalance" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.314575 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.314683 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj"] Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.316572 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.317258 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.467522 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-ring-data-devices\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.467581 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-swiftconf\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.467632 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbc46\" (UniqueName: \"kubernetes.io/projected/48e90b4f-5422-450f-9429-81000c29e36f-kube-api-access-cbc46\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.467670 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48e90b4f-5422-450f-9429-81000c29e36f-etc-swift\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.467689 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-dispersionconf\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.467902 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-scripts\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.569566 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbc46\" (UniqueName: \"kubernetes.io/projected/48e90b4f-5422-450f-9429-81000c29e36f-kube-api-access-cbc46\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.569648 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48e90b4f-5422-450f-9429-81000c29e36f-etc-swift\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.569675 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-dispersionconf\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.569715 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-scripts\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.569788 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-ring-data-devices\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.569831 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-swiftconf\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.570377 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48e90b4f-5422-450f-9429-81000c29e36f-etc-swift\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.570705 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-ring-data-devices\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.571281 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-scripts\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.577000 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-swiftconf\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.577376 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-dispersionconf\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.588722 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbc46\" (UniqueName: \"kubernetes.io/projected/48e90b4f-5422-450f-9429-81000c29e36f-kube-api-access-cbc46\") pod \"swift-ring-rebalance-debug-xt8hj\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.640464 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:52 crc kubenswrapper[4692]: I0309 09:54:52.929665 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj"] Mar 09 09:54:53 crc kubenswrapper[4692]: I0309 09:54:53.866622 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" event={"ID":"48e90b4f-5422-450f-9429-81000c29e36f","Type":"ContainerStarted","Data":"e5c2c85fde01a89a44f176830513b4333f8361e5afea337f991e5af9d8971c6e"} Mar 09 09:54:53 crc kubenswrapper[4692]: I0309 09:54:53.866999 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" event={"ID":"48e90b4f-5422-450f-9429-81000c29e36f","Type":"ContainerStarted","Data":"aa09462120090d4a877c40f27c9e6d505a12c5f8f25a9e2aaa1fc3abfccd67cd"} Mar 09 09:54:53 crc kubenswrapper[4692]: I0309 09:54:53.889133 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" podStartSLOduration=1.8891061279999999 podStartE2EDuration="1.889106128s" podCreationTimestamp="2026-03-09 09:54:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:53.883790578 +0000 UTC m=+2094.708526169" watchObservedRunningTime="2026-03-09 09:54:53.889106128 +0000 UTC m=+2094.713841709" Mar 09 09:54:54 crc kubenswrapper[4692]: I0309 09:54:54.881642 4692 generic.go:334] "Generic (PLEG): container finished" podID="48e90b4f-5422-450f-9429-81000c29e36f" containerID="e5c2c85fde01a89a44f176830513b4333f8361e5afea337f991e5af9d8971c6e" exitCode=0 Mar 09 09:54:54 crc kubenswrapper[4692]: I0309 09:54:54.881761 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" event={"ID":"48e90b4f-5422-450f-9429-81000c29e36f","Type":"ContainerDied","Data":"e5c2c85fde01a89a44f176830513b4333f8361e5afea337f991e5af9d8971c6e"} Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.172007 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.200008 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj"] Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.205689 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj"] Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.324211 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48e90b4f-5422-450f-9429-81000c29e36f-etc-swift\") pod \"48e90b4f-5422-450f-9429-81000c29e36f\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.324381 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-ring-data-devices\") pod \"48e90b4f-5422-450f-9429-81000c29e36f\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.324421 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbc46\" (UniqueName: \"kubernetes.io/projected/48e90b4f-5422-450f-9429-81000c29e36f-kube-api-access-cbc46\") pod \"48e90b4f-5422-450f-9429-81000c29e36f\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.324513 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-dispersionconf\") pod \"48e90b4f-5422-450f-9429-81000c29e36f\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.324543 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-scripts\") pod \"48e90b4f-5422-450f-9429-81000c29e36f\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.325210 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "48e90b4f-5422-450f-9429-81000c29e36f" (UID: "48e90b4f-5422-450f-9429-81000c29e36f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.325262 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48e90b4f-5422-450f-9429-81000c29e36f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "48e90b4f-5422-450f-9429-81000c29e36f" (UID: "48e90b4f-5422-450f-9429-81000c29e36f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.325280 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-swiftconf\") pod \"48e90b4f-5422-450f-9429-81000c29e36f\" (UID: \"48e90b4f-5422-450f-9429-81000c29e36f\") " Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.325660 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/48e90b4f-5422-450f-9429-81000c29e36f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.325682 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.330015 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e90b4f-5422-450f-9429-81000c29e36f-kube-api-access-cbc46" (OuterVolumeSpecName: "kube-api-access-cbc46") pod "48e90b4f-5422-450f-9429-81000c29e36f" (UID: "48e90b4f-5422-450f-9429-81000c29e36f"). InnerVolumeSpecName "kube-api-access-cbc46". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.348308 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-scripts" (OuterVolumeSpecName: "scripts") pod "48e90b4f-5422-450f-9429-81000c29e36f" (UID: "48e90b4f-5422-450f-9429-81000c29e36f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.355330 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "48e90b4f-5422-450f-9429-81000c29e36f" (UID: "48e90b4f-5422-450f-9429-81000c29e36f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.367468 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "48e90b4f-5422-450f-9429-81000c29e36f" (UID: "48e90b4f-5422-450f-9429-81000c29e36f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.427126 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.427156 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/48e90b4f-5422-450f-9429-81000c29e36f-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.427176 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/48e90b4f-5422-450f-9429-81000c29e36f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.427186 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbc46\" (UniqueName: \"kubernetes.io/projected/48e90b4f-5422-450f-9429-81000c29e36f-kube-api-access-cbc46\") on node \"crc\" DevicePath \"\"" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.898970 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa09462120090d4a877c40f27c9e6d505a12c5f8f25a9e2aaa1fc3abfccd67cd" Mar 09 09:54:56 crc kubenswrapper[4692]: I0309 09:54:56.899019 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xt8hj" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.355040 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q"] Mar 09 09:54:57 crc kubenswrapper[4692]: E0309 09:54:57.355434 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e90b4f-5422-450f-9429-81000c29e36f" containerName="swift-ring-rebalance" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.355448 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e90b4f-5422-450f-9429-81000c29e36f" containerName="swift-ring-rebalance" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.355655 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e90b4f-5422-450f-9429-81000c29e36f" containerName="swift-ring-rebalance" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.356186 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.358886 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.359299 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.362721 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q"] Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.544588 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr4bc\" (UniqueName: \"kubernetes.io/projected/e1173b40-88bb-4cbf-9516-13864406230c-kube-api-access-pr4bc\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.544956 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-dispersionconf\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.545032 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1173b40-88bb-4cbf-9516-13864406230c-etc-swift\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.545139 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-scripts\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.545203 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-swiftconf\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.545264 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-ring-data-devices\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.647127 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-dispersionconf\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.647193 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1173b40-88bb-4cbf-9516-13864406230c-etc-swift\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.647225 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-scripts\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.647242 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-swiftconf\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.647265 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-ring-data-devices\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.647285 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr4bc\" (UniqueName: \"kubernetes.io/projected/e1173b40-88bb-4cbf-9516-13864406230c-kube-api-access-pr4bc\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.648330 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1173b40-88bb-4cbf-9516-13864406230c-etc-swift\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.648369 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-ring-data-devices\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.648582 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-scripts\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.656642 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-dispersionconf\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.656681 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-swiftconf\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.662341 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr4bc\" (UniqueName: \"kubernetes.io/projected/e1173b40-88bb-4cbf-9516-13864406230c-kube-api-access-pr4bc\") pod \"swift-ring-rebalance-debug-6wm4q\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:57 crc kubenswrapper[4692]: I0309 09:54:57.717212 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:54:58 crc kubenswrapper[4692]: I0309 09:54:58.081125 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48e90b4f-5422-450f-9429-81000c29e36f" path="/var/lib/kubelet/pods/48e90b4f-5422-450f-9429-81000c29e36f/volumes" Mar 09 09:54:58 crc kubenswrapper[4692]: I0309 09:54:58.132083 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q"] Mar 09 09:54:58 crc kubenswrapper[4692]: I0309 09:54:58.918182 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" event={"ID":"e1173b40-88bb-4cbf-9516-13864406230c","Type":"ContainerStarted","Data":"0b60bfd273f7b234f723038ea0b4f37f5816c4ed4723ad914a7ce98f7a07678d"} Mar 09 09:54:58 crc kubenswrapper[4692]: I0309 09:54:58.918516 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" event={"ID":"e1173b40-88bb-4cbf-9516-13864406230c","Type":"ContainerStarted","Data":"45c9f57e8c8a44acdbae98dc9af74b45c970724e87f7833fa1be1d976eed5c5a"} Mar 09 09:54:58 crc kubenswrapper[4692]: I0309 09:54:58.939416 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" podStartSLOduration=1.939392871 podStartE2EDuration="1.939392871s" podCreationTimestamp="2026-03-09 09:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:54:58.933567226 +0000 UTC m=+2099.758302817" watchObservedRunningTime="2026-03-09 09:54:58.939392871 +0000 UTC m=+2099.764128462" Mar 09 09:54:59 crc kubenswrapper[4692]: I0309 09:54:59.926550 4692 generic.go:334] "Generic (PLEG): container finished" podID="e1173b40-88bb-4cbf-9516-13864406230c" containerID="0b60bfd273f7b234f723038ea0b4f37f5816c4ed4723ad914a7ce98f7a07678d" exitCode=0 Mar 09 09:54:59 crc kubenswrapper[4692]: I0309 09:54:59.926654 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" event={"ID":"e1173b40-88bb-4cbf-9516-13864406230c","Type":"ContainerDied","Data":"0b60bfd273f7b234f723038ea0b4f37f5816c4ed4723ad914a7ce98f7a07678d"} Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.226992 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.260030 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q"] Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.267822 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q"] Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424147 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-dispersionconf\") pod \"e1173b40-88bb-4cbf-9516-13864406230c\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424287 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-scripts\") pod \"e1173b40-88bb-4cbf-9516-13864406230c\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424324 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1173b40-88bb-4cbf-9516-13864406230c-etc-swift\") pod \"e1173b40-88bb-4cbf-9516-13864406230c\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424344 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-ring-data-devices\") pod \"e1173b40-88bb-4cbf-9516-13864406230c\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424374 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-swiftconf\") pod \"e1173b40-88bb-4cbf-9516-13864406230c\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424444 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr4bc\" (UniqueName: \"kubernetes.io/projected/e1173b40-88bb-4cbf-9516-13864406230c-kube-api-access-pr4bc\") pod \"e1173b40-88bb-4cbf-9516-13864406230c\" (UID: \"e1173b40-88bb-4cbf-9516-13864406230c\") " Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424773 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e1173b40-88bb-4cbf-9516-13864406230c" (UID: "e1173b40-88bb-4cbf-9516-13864406230c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.424986 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1173b40-88bb-4cbf-9516-13864406230c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e1173b40-88bb-4cbf-9516-13864406230c" (UID: "e1173b40-88bb-4cbf-9516-13864406230c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.429687 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1173b40-88bb-4cbf-9516-13864406230c-kube-api-access-pr4bc" (OuterVolumeSpecName: "kube-api-access-pr4bc") pod "e1173b40-88bb-4cbf-9516-13864406230c" (UID: "e1173b40-88bb-4cbf-9516-13864406230c"). InnerVolumeSpecName "kube-api-access-pr4bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.446801 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-scripts" (OuterVolumeSpecName: "scripts") pod "e1173b40-88bb-4cbf-9516-13864406230c" (UID: "e1173b40-88bb-4cbf-9516-13864406230c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.447406 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e1173b40-88bb-4cbf-9516-13864406230c" (UID: "e1173b40-88bb-4cbf-9516-13864406230c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.447458 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e1173b40-88bb-4cbf-9516-13864406230c" (UID: "e1173b40-88bb-4cbf-9516-13864406230c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.526886 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e1173b40-88bb-4cbf-9516-13864406230c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.527042 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.527065 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.527081 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr4bc\" (UniqueName: \"kubernetes.io/projected/e1173b40-88bb-4cbf-9516-13864406230c-kube-api-access-pr4bc\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.527095 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e1173b40-88bb-4cbf-9516-13864406230c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.527113 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e1173b40-88bb-4cbf-9516-13864406230c-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.941943 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c9f57e8c8a44acdbae98dc9af74b45c970724e87f7833fa1be1d976eed5c5a" Mar 09 09:55:01 crc kubenswrapper[4692]: I0309 09:55:01.942032 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6wm4q" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.079101 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1173b40-88bb-4cbf-9516-13864406230c" path="/var/lib/kubelet/pods/e1173b40-88bb-4cbf-9516-13864406230c/volumes" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.423309 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg"] Mar 09 09:55:02 crc kubenswrapper[4692]: E0309 09:55:02.424016 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1173b40-88bb-4cbf-9516-13864406230c" containerName="swift-ring-rebalance" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.424057 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1173b40-88bb-4cbf-9516-13864406230c" containerName="swift-ring-rebalance" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.424220 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1173b40-88bb-4cbf-9516-13864406230c" containerName="swift-ring-rebalance" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.424746 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.426719 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.426904 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.441468 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg"] Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.540672 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q76nq\" (UniqueName: \"kubernetes.io/projected/892cb975-4d1d-4c14-a939-f67d87c4a7b5-kube-api-access-q76nq\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.540780 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/892cb975-4d1d-4c14-a939-f67d87c4a7b5-etc-swift\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.540829 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-scripts\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.540847 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-dispersionconf\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.540863 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-ring-data-devices\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.540959 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-swiftconf\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.644778 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q76nq\" (UniqueName: \"kubernetes.io/projected/892cb975-4d1d-4c14-a939-f67d87c4a7b5-kube-api-access-q76nq\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.644829 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/892cb975-4d1d-4c14-a939-f67d87c4a7b5-etc-swift\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.644889 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-scripts\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.644908 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-dispersionconf\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.644923 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-ring-data-devices\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.644957 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-swiftconf\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.646033 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-scripts\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.646361 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/892cb975-4d1d-4c14-a939-f67d87c4a7b5-etc-swift\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.646772 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-ring-data-devices\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.650348 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-dispersionconf\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.653476 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-swiftconf\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.663908 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q76nq\" (UniqueName: \"kubernetes.io/projected/892cb975-4d1d-4c14-a939-f67d87c4a7b5-kube-api-access-q76nq\") pod \"swift-ring-rebalance-debug-6s4lg\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:02 crc kubenswrapper[4692]: I0309 09:55:02.745056 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:03 crc kubenswrapper[4692]: W0309 09:55:03.182212 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod892cb975_4d1d_4c14_a939_f67d87c4a7b5.slice/crio-29709fc1b6b80ccfb921d6a5e4520e6a73a68fa865558eb1e4794c1d16c02ada WatchSource:0}: Error finding container 29709fc1b6b80ccfb921d6a5e4520e6a73a68fa865558eb1e4794c1d16c02ada: Status 404 returned error can't find the container with id 29709fc1b6b80ccfb921d6a5e4520e6a73a68fa865558eb1e4794c1d16c02ada Mar 09 09:55:03 crc kubenswrapper[4692]: I0309 09:55:03.185323 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg"] Mar 09 09:55:03 crc kubenswrapper[4692]: I0309 09:55:03.961845 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" event={"ID":"892cb975-4d1d-4c14-a939-f67d87c4a7b5","Type":"ContainerStarted","Data":"828b3f9bb0c31f366f65b0271e5e0710357d7d5c49125d03befa7f534fd78afa"} Mar 09 09:55:03 crc kubenswrapper[4692]: I0309 09:55:03.962188 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" event={"ID":"892cb975-4d1d-4c14-a939-f67d87c4a7b5","Type":"ContainerStarted","Data":"29709fc1b6b80ccfb921d6a5e4520e6a73a68fa865558eb1e4794c1d16c02ada"} Mar 09 09:55:03 crc kubenswrapper[4692]: I0309 09:55:03.983982 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" podStartSLOduration=1.983959872 podStartE2EDuration="1.983959872s" podCreationTimestamp="2026-03-09 09:55:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:03.976741068 +0000 UTC m=+2104.801476659" watchObservedRunningTime="2026-03-09 09:55:03.983959872 +0000 UTC m=+2104.808695473" Mar 09 09:55:04 crc kubenswrapper[4692]: I0309 09:55:04.227976 4692 scope.go:117] "RemoveContainer" containerID="54d9bba991c4580b8ebc695ea147000337f9dc690409245583b6f21349e8c8b7" Mar 09 09:55:04 crc kubenswrapper[4692]: I0309 09:55:04.972944 4692 generic.go:334] "Generic (PLEG): container finished" podID="892cb975-4d1d-4c14-a939-f67d87c4a7b5" containerID="828b3f9bb0c31f366f65b0271e5e0710357d7d5c49125d03befa7f534fd78afa" exitCode=0 Mar 09 09:55:04 crc kubenswrapper[4692]: I0309 09:55:04.973113 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" event={"ID":"892cb975-4d1d-4c14-a939-f67d87c4a7b5","Type":"ContainerDied","Data":"828b3f9bb0c31f366f65b0271e5e0710357d7d5c49125d03befa7f534fd78afa"} Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.240367 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.267422 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg"] Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.274848 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg"] Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295019 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-scripts\") pod \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295093 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-dispersionconf\") pod \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295123 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q76nq\" (UniqueName: \"kubernetes.io/projected/892cb975-4d1d-4c14-a939-f67d87c4a7b5-kube-api-access-q76nq\") pod \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295149 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-ring-data-devices\") pod \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295177 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-swiftconf\") pod \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295232 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/892cb975-4d1d-4c14-a939-f67d87c4a7b5-etc-swift\") pod \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\" (UID: \"892cb975-4d1d-4c14-a939-f67d87c4a7b5\") " Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295684 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "892cb975-4d1d-4c14-a939-f67d87c4a7b5" (UID: "892cb975-4d1d-4c14-a939-f67d87c4a7b5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.295981 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/892cb975-4d1d-4c14-a939-f67d87c4a7b5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "892cb975-4d1d-4c14-a939-f67d87c4a7b5" (UID: "892cb975-4d1d-4c14-a939-f67d87c4a7b5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.300623 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/892cb975-4d1d-4c14-a939-f67d87c4a7b5-kube-api-access-q76nq" (OuterVolumeSpecName: "kube-api-access-q76nq") pod "892cb975-4d1d-4c14-a939-f67d87c4a7b5" (UID: "892cb975-4d1d-4c14-a939-f67d87c4a7b5"). InnerVolumeSpecName "kube-api-access-q76nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.316652 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-scripts" (OuterVolumeSpecName: "scripts") pod "892cb975-4d1d-4c14-a939-f67d87c4a7b5" (UID: "892cb975-4d1d-4c14-a939-f67d87c4a7b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.317383 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "892cb975-4d1d-4c14-a939-f67d87c4a7b5" (UID: "892cb975-4d1d-4c14-a939-f67d87c4a7b5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.318352 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "892cb975-4d1d-4c14-a939-f67d87c4a7b5" (UID: "892cb975-4d1d-4c14-a939-f67d87c4a7b5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.396571 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.396903 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.396922 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q76nq\" (UniqueName: \"kubernetes.io/projected/892cb975-4d1d-4c14-a939-f67d87c4a7b5-kube-api-access-q76nq\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.396934 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/892cb975-4d1d-4c14-a939-f67d87c4a7b5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.396946 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/892cb975-4d1d-4c14-a939-f67d87c4a7b5-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.396959 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/892cb975-4d1d-4c14-a939-f67d87c4a7b5-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.990226 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29709fc1b6b80ccfb921d6a5e4520e6a73a68fa865558eb1e4794c1d16c02ada" Mar 09 09:55:06 crc kubenswrapper[4692]: I0309 09:55:06.990270 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6s4lg" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.420904 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fztls"] Mar 09 09:55:07 crc kubenswrapper[4692]: E0309 09:55:07.421177 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="892cb975-4d1d-4c14-a939-f67d87c4a7b5" containerName="swift-ring-rebalance" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.421195 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="892cb975-4d1d-4c14-a939-f67d87c4a7b5" containerName="swift-ring-rebalance" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.421348 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="892cb975-4d1d-4c14-a939-f67d87c4a7b5" containerName="swift-ring-rebalance" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.421780 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.425357 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.425991 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.441638 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fztls"] Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.610575 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-dispersionconf\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.610836 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vlw2\" (UniqueName: \"kubernetes.io/projected/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-kube-api-access-9vlw2\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.610937 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-swiftconf\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.610986 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-ring-data-devices\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.611007 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-scripts\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.611097 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-etc-swift\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.711903 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-ring-data-devices\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.711956 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-scripts\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.711983 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-etc-swift\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.712011 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-dispersionconf\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.712042 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vlw2\" (UniqueName: \"kubernetes.io/projected/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-kube-api-access-9vlw2\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.712091 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-swiftconf\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.712699 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-etc-swift\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.712839 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-ring-data-devices\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.712936 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-scripts\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.717569 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-dispersionconf\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.717646 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-swiftconf\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.729388 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vlw2\" (UniqueName: \"kubernetes.io/projected/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-kube-api-access-9vlw2\") pod \"swift-ring-rebalance-debug-fztls\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:07 crc kubenswrapper[4692]: I0309 09:55:07.739300 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:08 crc kubenswrapper[4692]: I0309 09:55:08.081582 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="892cb975-4d1d-4c14-a939-f67d87c4a7b5" path="/var/lib/kubelet/pods/892cb975-4d1d-4c14-a939-f67d87c4a7b5/volumes" Mar 09 09:55:08 crc kubenswrapper[4692]: I0309 09:55:08.172307 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fztls"] Mar 09 09:55:08 crc kubenswrapper[4692]: W0309 09:55:08.176304 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod848e6a86_2ca8_44b5_9e7c_2d42d2e37310.slice/crio-61187c66d9048df1d168ee0a7165e9b97f316184b67008c23917328c281ef9ca WatchSource:0}: Error finding container 61187c66d9048df1d168ee0a7165e9b97f316184b67008c23917328c281ef9ca: Status 404 returned error can't find the container with id 61187c66d9048df1d168ee0a7165e9b97f316184b67008c23917328c281ef9ca Mar 09 09:55:09 crc kubenswrapper[4692]: I0309 09:55:09.005379 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" event={"ID":"848e6a86-2ca8-44b5-9e7c-2d42d2e37310","Type":"ContainerStarted","Data":"ba9798dc7eb1d2d2762cdb3f56cc3dabb328567811fe9aa71d948ce046363af8"} Mar 09 09:55:09 crc kubenswrapper[4692]: I0309 09:55:09.005871 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" event={"ID":"848e6a86-2ca8-44b5-9e7c-2d42d2e37310","Type":"ContainerStarted","Data":"61187c66d9048df1d168ee0a7165e9b97f316184b67008c23917328c281ef9ca"} Mar 09 09:55:09 crc kubenswrapper[4692]: I0309 09:55:09.023571 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" podStartSLOduration=2.023550862 podStartE2EDuration="2.023550862s" podCreationTimestamp="2026-03-09 09:55:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:09.021142684 +0000 UTC m=+2109.845878295" watchObservedRunningTime="2026-03-09 09:55:09.023550862 +0000 UTC m=+2109.848286443" Mar 09 09:55:10 crc kubenswrapper[4692]: I0309 09:55:10.016468 4692 generic.go:334] "Generic (PLEG): container finished" podID="848e6a86-2ca8-44b5-9e7c-2d42d2e37310" containerID="ba9798dc7eb1d2d2762cdb3f56cc3dabb328567811fe9aa71d948ce046363af8" exitCode=0 Mar 09 09:55:10 crc kubenswrapper[4692]: I0309 09:55:10.016532 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" event={"ID":"848e6a86-2ca8-44b5-9e7c-2d42d2e37310","Type":"ContainerDied","Data":"ba9798dc7eb1d2d2762cdb3f56cc3dabb328567811fe9aa71d948ce046363af8"} Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.352393 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.388690 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fztls"] Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.394175 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fztls"] Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.461418 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vlw2\" (UniqueName: \"kubernetes.io/projected/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-kube-api-access-9vlw2\") pod \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.461467 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-ring-data-devices\") pod \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.461507 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-etc-swift\") pod \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.461543 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-scripts\") pod \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.461563 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-swiftconf\") pod \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.461585 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-dispersionconf\") pod \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\" (UID: \"848e6a86-2ca8-44b5-9e7c-2d42d2e37310\") " Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.462475 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "848e6a86-2ca8-44b5-9e7c-2d42d2e37310" (UID: "848e6a86-2ca8-44b5-9e7c-2d42d2e37310"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.463034 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "848e6a86-2ca8-44b5-9e7c-2d42d2e37310" (UID: "848e6a86-2ca8-44b5-9e7c-2d42d2e37310"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.467102 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-kube-api-access-9vlw2" (OuterVolumeSpecName: "kube-api-access-9vlw2") pod "848e6a86-2ca8-44b5-9e7c-2d42d2e37310" (UID: "848e6a86-2ca8-44b5-9e7c-2d42d2e37310"). InnerVolumeSpecName "kube-api-access-9vlw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.484505 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-scripts" (OuterVolumeSpecName: "scripts") pod "848e6a86-2ca8-44b5-9e7c-2d42d2e37310" (UID: "848e6a86-2ca8-44b5-9e7c-2d42d2e37310"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.487109 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "848e6a86-2ca8-44b5-9e7c-2d42d2e37310" (UID: "848e6a86-2ca8-44b5-9e7c-2d42d2e37310"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.495652 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "848e6a86-2ca8-44b5-9e7c-2d42d2e37310" (UID: "848e6a86-2ca8-44b5-9e7c-2d42d2e37310"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.562684 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.562726 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vlw2\" (UniqueName: \"kubernetes.io/projected/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-kube-api-access-9vlw2\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.562755 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.562766 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.562776 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:11 crc kubenswrapper[4692]: I0309 09:55:11.562783 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/848e6a86-2ca8-44b5-9e7c-2d42d2e37310-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.034567 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61187c66d9048df1d168ee0a7165e9b97f316184b67008c23917328c281ef9ca" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.034610 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fztls" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.081417 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="848e6a86-2ca8-44b5-9e7c-2d42d2e37310" path="/var/lib/kubelet/pods/848e6a86-2ca8-44b5-9e7c-2d42d2e37310/volumes" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.539344 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk"] Mar 09 09:55:12 crc kubenswrapper[4692]: E0309 09:55:12.539693 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="848e6a86-2ca8-44b5-9e7c-2d42d2e37310" containerName="swift-ring-rebalance" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.539709 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="848e6a86-2ca8-44b5-9e7c-2d42d2e37310" containerName="swift-ring-rebalance" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.539891 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="848e6a86-2ca8-44b5-9e7c-2d42d2e37310" containerName="swift-ring-rebalance" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.540498 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.547884 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk"] Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.549001 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.549042 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.675948 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-dispersionconf\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.676281 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdd2l\" (UniqueName: \"kubernetes.io/projected/1ae60959-9000-415a-a8c9-04b5b1e72e37-kube-api-access-vdd2l\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.676379 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-scripts\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.676477 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-swiftconf\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.676637 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-ring-data-devices\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.676689 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1ae60959-9000-415a-a8c9-04b5b1e72e37-etc-swift\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.777896 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdd2l\" (UniqueName: \"kubernetes.io/projected/1ae60959-9000-415a-a8c9-04b5b1e72e37-kube-api-access-vdd2l\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.778253 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-scripts\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.778396 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-swiftconf\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.778609 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-ring-data-devices\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.778738 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1ae60959-9000-415a-a8c9-04b5b1e72e37-etc-swift\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.778893 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-dispersionconf\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.779051 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-scripts\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.779111 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-ring-data-devices\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.779251 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1ae60959-9000-415a-a8c9-04b5b1e72e37-etc-swift\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.784798 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-dispersionconf\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.788616 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-swiftconf\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.806287 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdd2l\" (UniqueName: \"kubernetes.io/projected/1ae60959-9000-415a-a8c9-04b5b1e72e37-kube-api-access-vdd2l\") pod \"swift-ring-rebalance-debug-s9cwk\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:12 crc kubenswrapper[4692]: I0309 09:55:12.862979 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:13 crc kubenswrapper[4692]: I0309 09:55:13.339873 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk"] Mar 09 09:55:14 crc kubenswrapper[4692]: I0309 09:55:14.057364 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" event={"ID":"1ae60959-9000-415a-a8c9-04b5b1e72e37","Type":"ContainerStarted","Data":"c9f7949195aeeb4133ea50e703e4d65a96217257e14b663b9e7f27e475897099"} Mar 09 09:55:14 crc kubenswrapper[4692]: I0309 09:55:14.057729 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" event={"ID":"1ae60959-9000-415a-a8c9-04b5b1e72e37","Type":"ContainerStarted","Data":"7d5b48ff99715544b4a5eb5e48fe265bc2f52953452b529bd3210e6aa49fdef0"} Mar 09 09:55:14 crc kubenswrapper[4692]: I0309 09:55:14.082481 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" podStartSLOduration=2.082464388 podStartE2EDuration="2.082464388s" podCreationTimestamp="2026-03-09 09:55:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:14.074541544 +0000 UTC m=+2114.899277125" watchObservedRunningTime="2026-03-09 09:55:14.082464388 +0000 UTC m=+2114.907199959" Mar 09 09:55:15 crc kubenswrapper[4692]: I0309 09:55:15.066098 4692 generic.go:334] "Generic (PLEG): container finished" podID="1ae60959-9000-415a-a8c9-04b5b1e72e37" containerID="c9f7949195aeeb4133ea50e703e4d65a96217257e14b663b9e7f27e475897099" exitCode=0 Mar 09 09:55:15 crc kubenswrapper[4692]: I0309 09:55:15.066216 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" event={"ID":"1ae60959-9000-415a-a8c9-04b5b1e72e37","Type":"ContainerDied","Data":"c9f7949195aeeb4133ea50e703e4d65a96217257e14b663b9e7f27e475897099"} Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.376962 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.404198 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk"] Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.410814 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk"] Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.527513 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-dispersionconf\") pod \"1ae60959-9000-415a-a8c9-04b5b1e72e37\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.527638 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1ae60959-9000-415a-a8c9-04b5b1e72e37-etc-swift\") pod \"1ae60959-9000-415a-a8c9-04b5b1e72e37\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.527698 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdd2l\" (UniqueName: \"kubernetes.io/projected/1ae60959-9000-415a-a8c9-04b5b1e72e37-kube-api-access-vdd2l\") pod \"1ae60959-9000-415a-a8c9-04b5b1e72e37\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.527744 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-scripts\") pod \"1ae60959-9000-415a-a8c9-04b5b1e72e37\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.527770 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-swiftconf\") pod \"1ae60959-9000-415a-a8c9-04b5b1e72e37\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.527830 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-ring-data-devices\") pod \"1ae60959-9000-415a-a8c9-04b5b1e72e37\" (UID: \"1ae60959-9000-415a-a8c9-04b5b1e72e37\") " Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.528557 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ae60959-9000-415a-a8c9-04b5b1e72e37-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1ae60959-9000-415a-a8c9-04b5b1e72e37" (UID: "1ae60959-9000-415a-a8c9-04b5b1e72e37"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.528682 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1ae60959-9000-415a-a8c9-04b5b1e72e37" (UID: "1ae60959-9000-415a-a8c9-04b5b1e72e37"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.535264 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ae60959-9000-415a-a8c9-04b5b1e72e37-kube-api-access-vdd2l" (OuterVolumeSpecName: "kube-api-access-vdd2l") pod "1ae60959-9000-415a-a8c9-04b5b1e72e37" (UID: "1ae60959-9000-415a-a8c9-04b5b1e72e37"). InnerVolumeSpecName "kube-api-access-vdd2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.546753 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-scripts" (OuterVolumeSpecName: "scripts") pod "1ae60959-9000-415a-a8c9-04b5b1e72e37" (UID: "1ae60959-9000-415a-a8c9-04b5b1e72e37"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.552397 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1ae60959-9000-415a-a8c9-04b5b1e72e37" (UID: "1ae60959-9000-415a-a8c9-04b5b1e72e37"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.557348 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1ae60959-9000-415a-a8c9-04b5b1e72e37" (UID: "1ae60959-9000-415a-a8c9-04b5b1e72e37"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.629905 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.629953 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.629966 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1ae60959-9000-415a-a8c9-04b5b1e72e37-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.629979 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdd2l\" (UniqueName: \"kubernetes.io/projected/1ae60959-9000-415a-a8c9-04b5b1e72e37-kube-api-access-vdd2l\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.629992 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ae60959-9000-415a-a8c9-04b5b1e72e37-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:16 crc kubenswrapper[4692]: I0309 09:55:16.630003 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1ae60959-9000-415a-a8c9-04b5b1e72e37-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.085122 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d5b48ff99715544b4a5eb5e48fe265bc2f52953452b529bd3210e6aa49fdef0" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.085193 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-s9cwk" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.560612 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn"] Mar 09 09:55:17 crc kubenswrapper[4692]: E0309 09:55:17.561282 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ae60959-9000-415a-a8c9-04b5b1e72e37" containerName="swift-ring-rebalance" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.561300 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ae60959-9000-415a-a8c9-04b5b1e72e37" containerName="swift-ring-rebalance" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.561458 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ae60959-9000-415a-a8c9-04b5b1e72e37" containerName="swift-ring-rebalance" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.562047 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.564779 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.564882 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.571855 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn"] Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.748018 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-swiftconf\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.748066 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-dispersionconf\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.748113 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-etc-swift\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.748417 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-scripts\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.748511 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sb92\" (UniqueName: \"kubernetes.io/projected/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-kube-api-access-4sb92\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.748549 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-ring-data-devices\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.849355 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-swiftconf\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.849414 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-dispersionconf\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.849453 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-etc-swift\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.849502 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-scripts\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.849524 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sb92\" (UniqueName: \"kubernetes.io/projected/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-kube-api-access-4sb92\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.849545 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-ring-data-devices\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.850201 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-etc-swift\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.850336 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-ring-data-devices\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.850585 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-scripts\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.854511 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-dispersionconf\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.854557 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-swiftconf\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.868147 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sb92\" (UniqueName: \"kubernetes.io/projected/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-kube-api-access-4sb92\") pod \"swift-ring-rebalance-debug-bfsgn\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:17 crc kubenswrapper[4692]: I0309 09:55:17.881130 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:18 crc kubenswrapper[4692]: I0309 09:55:18.086501 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ae60959-9000-415a-a8c9-04b5b1e72e37" path="/var/lib/kubelet/pods/1ae60959-9000-415a-a8c9-04b5b1e72e37/volumes" Mar 09 09:55:18 crc kubenswrapper[4692]: I0309 09:55:18.275701 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn"] Mar 09 09:55:19 crc kubenswrapper[4692]: I0309 09:55:19.103352 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" event={"ID":"71b0e734-ccd2-4c97-940a-f537f1e6e1e0","Type":"ContainerStarted","Data":"a3cdb3789fd01824c1f6c40086934150a2b32ee0113798bc3ebab258324afc68"} Mar 09 09:55:19 crc kubenswrapper[4692]: I0309 09:55:19.103711 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" event={"ID":"71b0e734-ccd2-4c97-940a-f537f1e6e1e0","Type":"ContainerStarted","Data":"8bd9ff351e06ec27c43c5fc25f00ab3a4ff68fede4a331211aee3da7cd57130d"} Mar 09 09:55:19 crc kubenswrapper[4692]: I0309 09:55:19.122979 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" podStartSLOduration=2.122960414 podStartE2EDuration="2.122960414s" podCreationTimestamp="2026-03-09 09:55:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:19.122402658 +0000 UTC m=+2119.947138249" watchObservedRunningTime="2026-03-09 09:55:19.122960414 +0000 UTC m=+2119.947695995" Mar 09 09:55:20 crc kubenswrapper[4692]: I0309 09:55:20.112056 4692 generic.go:334] "Generic (PLEG): container finished" podID="71b0e734-ccd2-4c97-940a-f537f1e6e1e0" containerID="a3cdb3789fd01824c1f6c40086934150a2b32ee0113798bc3ebab258324afc68" exitCode=0 Mar 09 09:55:20 crc kubenswrapper[4692]: I0309 09:55:20.112151 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" event={"ID":"71b0e734-ccd2-4c97-940a-f537f1e6e1e0","Type":"ContainerDied","Data":"a3cdb3789fd01824c1f6c40086934150a2b32ee0113798bc3ebab258324afc68"} Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.433582 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.465303 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn"] Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.470190 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn"] Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.609883 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-dispersionconf\") pod \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.609957 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-swiftconf\") pod \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.609980 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-scripts\") pod \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.610080 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-etc-swift\") pod \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.610103 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sb92\" (UniqueName: \"kubernetes.io/projected/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-kube-api-access-4sb92\") pod \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.610148 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-ring-data-devices\") pod \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\" (UID: \"71b0e734-ccd2-4c97-940a-f537f1e6e1e0\") " Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.611059 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "71b0e734-ccd2-4c97-940a-f537f1e6e1e0" (UID: "71b0e734-ccd2-4c97-940a-f537f1e6e1e0"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.611114 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "71b0e734-ccd2-4c97-940a-f537f1e6e1e0" (UID: "71b0e734-ccd2-4c97-940a-f537f1e6e1e0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.615509 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-kube-api-access-4sb92" (OuterVolumeSpecName: "kube-api-access-4sb92") pod "71b0e734-ccd2-4c97-940a-f537f1e6e1e0" (UID: "71b0e734-ccd2-4c97-940a-f537f1e6e1e0"). InnerVolumeSpecName "kube-api-access-4sb92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.629019 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-scripts" (OuterVolumeSpecName: "scripts") pod "71b0e734-ccd2-4c97-940a-f537f1e6e1e0" (UID: "71b0e734-ccd2-4c97-940a-f537f1e6e1e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.631059 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "71b0e734-ccd2-4c97-940a-f537f1e6e1e0" (UID: "71b0e734-ccd2-4c97-940a-f537f1e6e1e0"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.632039 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "71b0e734-ccd2-4c97-940a-f537f1e6e1e0" (UID: "71b0e734-ccd2-4c97-940a-f537f1e6e1e0"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.712297 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.712333 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.712350 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.712362 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.712373 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:21 crc kubenswrapper[4692]: I0309 09:55:21.712388 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sb92\" (UniqueName: \"kubernetes.io/projected/71b0e734-ccd2-4c97-940a-f537f1e6e1e0-kube-api-access-4sb92\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.085808 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71b0e734-ccd2-4c97-940a-f537f1e6e1e0" path="/var/lib/kubelet/pods/71b0e734-ccd2-4c97-940a-f537f1e6e1e0/volumes" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.129748 4692 scope.go:117] "RemoveContainer" containerID="a3cdb3789fd01824c1f6c40086934150a2b32ee0113798bc3ebab258324afc68" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.129790 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bfsgn" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.606786 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq"] Mar 09 09:55:22 crc kubenswrapper[4692]: E0309 09:55:22.607186 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71b0e734-ccd2-4c97-940a-f537f1e6e1e0" containerName="swift-ring-rebalance" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.607203 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="71b0e734-ccd2-4c97-940a-f537f1e6e1e0" containerName="swift-ring-rebalance" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.607380 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="71b0e734-ccd2-4c97-940a-f537f1e6e1e0" containerName="swift-ring-rebalance" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.607995 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.611416 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.611605 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.629445 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq"] Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.725266 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-scripts\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.725365 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c679d03-43a7-4918-a7e3-1bd047920381-etc-swift\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.725416 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-dispersionconf\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.725449 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-ring-data-devices\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.725470 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-swiftconf\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.725516 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz2tx\" (UniqueName: \"kubernetes.io/projected/2c679d03-43a7-4918-a7e3-1bd047920381-kube-api-access-wz2tx\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.826936 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-scripts\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.826986 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c679d03-43a7-4918-a7e3-1bd047920381-etc-swift\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.827024 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-dispersionconf\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.827051 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-ring-data-devices\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.827070 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-swiftconf\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.827091 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz2tx\" (UniqueName: \"kubernetes.io/projected/2c679d03-43a7-4918-a7e3-1bd047920381-kube-api-access-wz2tx\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.827707 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c679d03-43a7-4918-a7e3-1bd047920381-etc-swift\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.827758 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-scripts\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.827860 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-ring-data-devices\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.831203 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-swiftconf\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.833488 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-dispersionconf\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.842578 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz2tx\" (UniqueName: \"kubernetes.io/projected/2c679d03-43a7-4918-a7e3-1bd047920381-kube-api-access-wz2tx\") pod \"swift-ring-rebalance-debug-xfpgq\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:22 crc kubenswrapper[4692]: I0309 09:55:22.926867 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:23 crc kubenswrapper[4692]: I0309 09:55:23.332043 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq"] Mar 09 09:55:23 crc kubenswrapper[4692]: W0309 09:55:23.335429 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c679d03_43a7_4918_a7e3_1bd047920381.slice/crio-1dcf8a0a37e6fd0383c047bd496c938d396064a249deb637a1cd9c4843758d67 WatchSource:0}: Error finding container 1dcf8a0a37e6fd0383c047bd496c938d396064a249deb637a1cd9c4843758d67: Status 404 returned error can't find the container with id 1dcf8a0a37e6fd0383c047bd496c938d396064a249deb637a1cd9c4843758d67 Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.157721 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" event={"ID":"2c679d03-43a7-4918-a7e3-1bd047920381","Type":"ContainerStarted","Data":"465939265c687ecb225c88675a6dc7746b4fa99f66577d802a7266bed10040ef"} Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.158190 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" event={"ID":"2c679d03-43a7-4918-a7e3-1bd047920381","Type":"ContainerStarted","Data":"1dcf8a0a37e6fd0383c047bd496c938d396064a249deb637a1cd9c4843758d67"} Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.192309 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" podStartSLOduration=2.192281036 podStartE2EDuration="2.192281036s" podCreationTimestamp="2026-03-09 09:55:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:24.183884598 +0000 UTC m=+2125.008620179" watchObservedRunningTime="2026-03-09 09:55:24.192281036 +0000 UTC m=+2125.017016617" Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.784783 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-28tcl"] Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.786589 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.796828 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-28tcl"] Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.958886 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86wc8\" (UniqueName: \"kubernetes.io/projected/83338b51-8ebd-45e5-bee1-d20b036c37fc-kube-api-access-86wc8\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.958941 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-utilities\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:24 crc kubenswrapper[4692]: I0309 09:55:24.959035 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-catalog-content\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.078452 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86wc8\" (UniqueName: \"kubernetes.io/projected/83338b51-8ebd-45e5-bee1-d20b036c37fc-kube-api-access-86wc8\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.078588 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-utilities\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.078908 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-catalog-content\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.080747 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-utilities\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.080873 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-catalog-content\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.102474 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86wc8\" (UniqueName: \"kubernetes.io/projected/83338b51-8ebd-45e5-bee1-d20b036c37fc-kube-api-access-86wc8\") pod \"community-operators-28tcl\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.135308 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.186319 4692 generic.go:334] "Generic (PLEG): container finished" podID="2c679d03-43a7-4918-a7e3-1bd047920381" containerID="465939265c687ecb225c88675a6dc7746b4fa99f66577d802a7266bed10040ef" exitCode=0 Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.186375 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" event={"ID":"2c679d03-43a7-4918-a7e3-1bd047920381","Type":"ContainerDied","Data":"465939265c687ecb225c88675a6dc7746b4fa99f66577d802a7266bed10040ef"} Mar 09 09:55:25 crc kubenswrapper[4692]: I0309 09:55:25.706707 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-28tcl"] Mar 09 09:55:25 crc kubenswrapper[4692]: W0309 09:55:25.713280 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83338b51_8ebd_45e5_bee1_d20b036c37fc.slice/crio-2dd4ae4a4ee4b8cbde7ab762353c9e3c45b405f297b1ac2b9140e125ef8b4a3c WatchSource:0}: Error finding container 2dd4ae4a4ee4b8cbde7ab762353c9e3c45b405f297b1ac2b9140e125ef8b4a3c: Status 404 returned error can't find the container with id 2dd4ae4a4ee4b8cbde7ab762353c9e3c45b405f297b1ac2b9140e125ef8b4a3c Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.195852 4692 generic.go:334] "Generic (PLEG): container finished" podID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerID="65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2" exitCode=0 Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.195910 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28tcl" event={"ID":"83338b51-8ebd-45e5-bee1-d20b036c37fc","Type":"ContainerDied","Data":"65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2"} Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.196323 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28tcl" event={"ID":"83338b51-8ebd-45e5-bee1-d20b036c37fc","Type":"ContainerStarted","Data":"2dd4ae4a4ee4b8cbde7ab762353c9e3c45b405f297b1ac2b9140e125ef8b4a3c"} Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.198032 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.486142 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.519063 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq"] Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.525702 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq"] Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.607691 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-ring-data-devices\") pod \"2c679d03-43a7-4918-a7e3-1bd047920381\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.607801 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c679d03-43a7-4918-a7e3-1bd047920381-etc-swift\") pod \"2c679d03-43a7-4918-a7e3-1bd047920381\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.607864 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-scripts\") pod \"2c679d03-43a7-4918-a7e3-1bd047920381\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.607889 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-dispersionconf\") pod \"2c679d03-43a7-4918-a7e3-1bd047920381\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.607953 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz2tx\" (UniqueName: \"kubernetes.io/projected/2c679d03-43a7-4918-a7e3-1bd047920381-kube-api-access-wz2tx\") pod \"2c679d03-43a7-4918-a7e3-1bd047920381\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.607979 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-swiftconf\") pod \"2c679d03-43a7-4918-a7e3-1bd047920381\" (UID: \"2c679d03-43a7-4918-a7e3-1bd047920381\") " Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.608642 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2c679d03-43a7-4918-a7e3-1bd047920381" (UID: "2c679d03-43a7-4918-a7e3-1bd047920381"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.608824 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c679d03-43a7-4918-a7e3-1bd047920381-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2c679d03-43a7-4918-a7e3-1bd047920381" (UID: "2c679d03-43a7-4918-a7e3-1bd047920381"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.608972 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2c679d03-43a7-4918-a7e3-1bd047920381-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.608996 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.613062 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c679d03-43a7-4918-a7e3-1bd047920381-kube-api-access-wz2tx" (OuterVolumeSpecName: "kube-api-access-wz2tx") pod "2c679d03-43a7-4918-a7e3-1bd047920381" (UID: "2c679d03-43a7-4918-a7e3-1bd047920381"). InnerVolumeSpecName "kube-api-access-wz2tx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.632083 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2c679d03-43a7-4918-a7e3-1bd047920381" (UID: "2c679d03-43a7-4918-a7e3-1bd047920381"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.638709 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-scripts" (OuterVolumeSpecName: "scripts") pod "2c679d03-43a7-4918-a7e3-1bd047920381" (UID: "2c679d03-43a7-4918-a7e3-1bd047920381"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.638875 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2c679d03-43a7-4918-a7e3-1bd047920381" (UID: "2c679d03-43a7-4918-a7e3-1bd047920381"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.709776 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz2tx\" (UniqueName: \"kubernetes.io/projected/2c679d03-43a7-4918-a7e3-1bd047920381-kube-api-access-wz2tx\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.709809 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.709821 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2c679d03-43a7-4918-a7e3-1bd047920381-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:26 crc kubenswrapper[4692]: I0309 09:55:26.709829 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2c679d03-43a7-4918-a7e3-1bd047920381-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.204804 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dcf8a0a37e6fd0383c047bd496c938d396064a249deb637a1cd9c4843758d67" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.204882 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xfpgq" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.667249 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k"] Mar 09 09:55:27 crc kubenswrapper[4692]: E0309 09:55:27.667605 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c679d03-43a7-4918-a7e3-1bd047920381" containerName="swift-ring-rebalance" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.667623 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c679d03-43a7-4918-a7e3-1bd047920381" containerName="swift-ring-rebalance" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.667770 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c679d03-43a7-4918-a7e3-1bd047920381" containerName="swift-ring-rebalance" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.668254 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.670306 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.682050 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.684145 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k"] Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.822952 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-scripts\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.823105 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-dispersionconf\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.823145 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkctg\" (UniqueName: \"kubernetes.io/projected/a6579068-a3f4-413b-9aed-fe1d0afd9733-kube-api-access-rkctg\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.823187 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-swiftconf\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.823302 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a6579068-a3f4-413b-9aed-fe1d0afd9733-etc-swift\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.823323 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-ring-data-devices\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.924187 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a6579068-a3f4-413b-9aed-fe1d0afd9733-etc-swift\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.924256 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-ring-data-devices\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.924300 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-scripts\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.924342 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-dispersionconf\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.924375 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkctg\" (UniqueName: \"kubernetes.io/projected/a6579068-a3f4-413b-9aed-fe1d0afd9733-kube-api-access-rkctg\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.924396 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-swiftconf\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.925364 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a6579068-a3f4-413b-9aed-fe1d0afd9733-etc-swift\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.925653 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-ring-data-devices\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.925868 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-scripts\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.929986 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-swiftconf\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.930446 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-dispersionconf\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.951459 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkctg\" (UniqueName: \"kubernetes.io/projected/a6579068-a3f4-413b-9aed-fe1d0afd9733-kube-api-access-rkctg\") pod \"swift-ring-rebalance-debug-kxh7k\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:27 crc kubenswrapper[4692]: I0309 09:55:27.984324 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:28 crc kubenswrapper[4692]: I0309 09:55:28.098050 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c679d03-43a7-4918-a7e3-1bd047920381" path="/var/lib/kubelet/pods/2c679d03-43a7-4918-a7e3-1bd047920381/volumes" Mar 09 09:55:28 crc kubenswrapper[4692]: I0309 09:55:28.214711 4692 generic.go:334] "Generic (PLEG): container finished" podID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerID="7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572" exitCode=0 Mar 09 09:55:28 crc kubenswrapper[4692]: I0309 09:55:28.214775 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28tcl" event={"ID":"83338b51-8ebd-45e5-bee1-d20b036c37fc","Type":"ContainerDied","Data":"7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572"} Mar 09 09:55:28 crc kubenswrapper[4692]: I0309 09:55:28.390550 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k"] Mar 09 09:55:28 crc kubenswrapper[4692]: W0309 09:55:28.396675 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6579068_a3f4_413b_9aed_fe1d0afd9733.slice/crio-988ff97701f91458f1a66c4ecaa961ce52bda2cec6427afbe899bf181d012c63 WatchSource:0}: Error finding container 988ff97701f91458f1a66c4ecaa961ce52bda2cec6427afbe899bf181d012c63: Status 404 returned error can't find the container with id 988ff97701f91458f1a66c4ecaa961ce52bda2cec6427afbe899bf181d012c63 Mar 09 09:55:29 crc kubenswrapper[4692]: I0309 09:55:29.229283 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28tcl" event={"ID":"83338b51-8ebd-45e5-bee1-d20b036c37fc","Type":"ContainerStarted","Data":"2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51"} Mar 09 09:55:29 crc kubenswrapper[4692]: I0309 09:55:29.232148 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" event={"ID":"a6579068-a3f4-413b-9aed-fe1d0afd9733","Type":"ContainerStarted","Data":"9f5573fdeb9526abf86dd80c7229949049acea80c0276a84dda1af12eb477ec8"} Mar 09 09:55:29 crc kubenswrapper[4692]: I0309 09:55:29.232233 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" event={"ID":"a6579068-a3f4-413b-9aed-fe1d0afd9733","Type":"ContainerStarted","Data":"988ff97701f91458f1a66c4ecaa961ce52bda2cec6427afbe899bf181d012c63"} Mar 09 09:55:29 crc kubenswrapper[4692]: I0309 09:55:29.255787 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-28tcl" podStartSLOduration=2.426873087 podStartE2EDuration="5.255766081s" podCreationTimestamp="2026-03-09 09:55:24 +0000 UTC" firstStartedPulling="2026-03-09 09:55:26.197761322 +0000 UTC m=+2127.022496903" lastFinishedPulling="2026-03-09 09:55:29.026654316 +0000 UTC m=+2129.851389897" observedRunningTime="2026-03-09 09:55:29.253120056 +0000 UTC m=+2130.077855657" watchObservedRunningTime="2026-03-09 09:55:29.255766081 +0000 UTC m=+2130.080501662" Mar 09 09:55:29 crc kubenswrapper[4692]: I0309 09:55:29.275927 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" podStartSLOduration=2.275903471 podStartE2EDuration="2.275903471s" podCreationTimestamp="2026-03-09 09:55:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:29.272115864 +0000 UTC m=+2130.096851475" watchObservedRunningTime="2026-03-09 09:55:29.275903471 +0000 UTC m=+2130.100639062" Mar 09 09:55:31 crc kubenswrapper[4692]: I0309 09:55:31.249802 4692 generic.go:334] "Generic (PLEG): container finished" podID="a6579068-a3f4-413b-9aed-fe1d0afd9733" containerID="9f5573fdeb9526abf86dd80c7229949049acea80c0276a84dda1af12eb477ec8" exitCode=0 Mar 09 09:55:31 crc kubenswrapper[4692]: I0309 09:55:31.249903 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" event={"ID":"a6579068-a3f4-413b-9aed-fe1d0afd9733","Type":"ContainerDied","Data":"9f5573fdeb9526abf86dd80c7229949049acea80c0276a84dda1af12eb477ec8"} Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.537240 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.570212 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k"] Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.577504 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k"] Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.713205 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a6579068-a3f4-413b-9aed-fe1d0afd9733-etc-swift\") pod \"a6579068-a3f4-413b-9aed-fe1d0afd9733\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.713281 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkctg\" (UniqueName: \"kubernetes.io/projected/a6579068-a3f4-413b-9aed-fe1d0afd9733-kube-api-access-rkctg\") pod \"a6579068-a3f4-413b-9aed-fe1d0afd9733\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.713323 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-dispersionconf\") pod \"a6579068-a3f4-413b-9aed-fe1d0afd9733\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.713350 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-scripts\") pod \"a6579068-a3f4-413b-9aed-fe1d0afd9733\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.713377 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-ring-data-devices\") pod \"a6579068-a3f4-413b-9aed-fe1d0afd9733\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.713436 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-swiftconf\") pod \"a6579068-a3f4-413b-9aed-fe1d0afd9733\" (UID: \"a6579068-a3f4-413b-9aed-fe1d0afd9733\") " Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.714051 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a6579068-a3f4-413b-9aed-fe1d0afd9733" (UID: "a6579068-a3f4-413b-9aed-fe1d0afd9733"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.714218 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6579068-a3f4-413b-9aed-fe1d0afd9733-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a6579068-a3f4-413b-9aed-fe1d0afd9733" (UID: "a6579068-a3f4-413b-9aed-fe1d0afd9733"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.719745 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6579068-a3f4-413b-9aed-fe1d0afd9733-kube-api-access-rkctg" (OuterVolumeSpecName: "kube-api-access-rkctg") pod "a6579068-a3f4-413b-9aed-fe1d0afd9733" (UID: "a6579068-a3f4-413b-9aed-fe1d0afd9733"). InnerVolumeSpecName "kube-api-access-rkctg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.736906 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-scripts" (OuterVolumeSpecName: "scripts") pod "a6579068-a3f4-413b-9aed-fe1d0afd9733" (UID: "a6579068-a3f4-413b-9aed-fe1d0afd9733"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.740855 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a6579068-a3f4-413b-9aed-fe1d0afd9733" (UID: "a6579068-a3f4-413b-9aed-fe1d0afd9733"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.740975 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a6579068-a3f4-413b-9aed-fe1d0afd9733" (UID: "a6579068-a3f4-413b-9aed-fe1d0afd9733"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.814588 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.814627 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a6579068-a3f4-413b-9aed-fe1d0afd9733-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.814641 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkctg\" (UniqueName: \"kubernetes.io/projected/a6579068-a3f4-413b-9aed-fe1d0afd9733-kube-api-access-rkctg\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.814654 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a6579068-a3f4-413b-9aed-fe1d0afd9733-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.814664 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:32 crc kubenswrapper[4692]: I0309 09:55:32.814673 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a6579068-a3f4-413b-9aed-fe1d0afd9733-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.266178 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="988ff97701f91458f1a66c4ecaa961ce52bda2cec6427afbe899bf181d012c63" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.266242 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kxh7k" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.734877 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4x6md"] Mar 09 09:55:33 crc kubenswrapper[4692]: E0309 09:55:33.735266 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6579068-a3f4-413b-9aed-fe1d0afd9733" containerName="swift-ring-rebalance" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.735282 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6579068-a3f4-413b-9aed-fe1d0afd9733" containerName="swift-ring-rebalance" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.735468 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6579068-a3f4-413b-9aed-fe1d0afd9733" containerName="swift-ring-rebalance" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.736083 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.737670 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.741673 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.743690 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4x6md"] Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.928748 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-ring-data-devices\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.928833 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-scripts\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.928854 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31ef19db-14de-48b2-b960-5c2dbffbe85d-etc-swift\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.928998 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-swiftconf\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.929022 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-dispersionconf\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:33 crc kubenswrapper[4692]: I0309 09:55:33.929047 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hgtg\" (UniqueName: \"kubernetes.io/projected/31ef19db-14de-48b2-b960-5c2dbffbe85d-kube-api-access-6hgtg\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.030498 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-swiftconf\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.030558 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-dispersionconf\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.030591 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hgtg\" (UniqueName: \"kubernetes.io/projected/31ef19db-14de-48b2-b960-5c2dbffbe85d-kube-api-access-6hgtg\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.030628 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-ring-data-devices\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.030680 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31ef19db-14de-48b2-b960-5c2dbffbe85d-etc-swift\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.030704 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-scripts\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.031380 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31ef19db-14de-48b2-b960-5c2dbffbe85d-etc-swift\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.031628 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-ring-data-devices\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.031628 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-scripts\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.034914 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-dispersionconf\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.038001 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-swiftconf\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.047669 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hgtg\" (UniqueName: \"kubernetes.io/projected/31ef19db-14de-48b2-b960-5c2dbffbe85d-kube-api-access-6hgtg\") pod \"swift-ring-rebalance-debug-4x6md\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.053543 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.080312 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6579068-a3f4-413b-9aed-fe1d0afd9733" path="/var/lib/kubelet/pods/a6579068-a3f4-413b-9aed-fe1d0afd9733/volumes" Mar 09 09:55:34 crc kubenswrapper[4692]: I0309 09:55:34.498626 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4x6md"] Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.135746 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.136173 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.186685 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.297772 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" event={"ID":"31ef19db-14de-48b2-b960-5c2dbffbe85d","Type":"ContainerStarted","Data":"5fe7772552024d4bf9decec7c373bee2a1eecc3ab2ae65b2d1687934a7109b34"} Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.297818 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" event={"ID":"31ef19db-14de-48b2-b960-5c2dbffbe85d","Type":"ContainerStarted","Data":"46a8eb6cd6402ce9a9e790d1fbde9090c73e8aeafa005f3f7eb882dce2f54240"} Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.321397 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" podStartSLOduration=2.321377755 podStartE2EDuration="2.321377755s" podCreationTimestamp="2026-03-09 09:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:35.315724014 +0000 UTC m=+2136.140459615" watchObservedRunningTime="2026-03-09 09:55:35.321377755 +0000 UTC m=+2136.146113346" Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.349747 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:35 crc kubenswrapper[4692]: I0309 09:55:35.423743 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-28tcl"] Mar 09 09:55:36 crc kubenswrapper[4692]: I0309 09:55:36.306414 4692 generic.go:334] "Generic (PLEG): container finished" podID="31ef19db-14de-48b2-b960-5c2dbffbe85d" containerID="5fe7772552024d4bf9decec7c373bee2a1eecc3ab2ae65b2d1687934a7109b34" exitCode=0 Mar 09 09:55:36 crc kubenswrapper[4692]: I0309 09:55:36.306488 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" event={"ID":"31ef19db-14de-48b2-b960-5c2dbffbe85d","Type":"ContainerDied","Data":"5fe7772552024d4bf9decec7c373bee2a1eecc3ab2ae65b2d1687934a7109b34"} Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.314507 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-28tcl" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="registry-server" containerID="cri-o://2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51" gracePeriod=2 Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.644617 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.682495 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4x6md"] Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.700212 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4x6md"] Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.797279 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hgtg\" (UniqueName: \"kubernetes.io/projected/31ef19db-14de-48b2-b960-5c2dbffbe85d-kube-api-access-6hgtg\") pod \"31ef19db-14de-48b2-b960-5c2dbffbe85d\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.797590 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-swiftconf\") pod \"31ef19db-14de-48b2-b960-5c2dbffbe85d\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.797768 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-ring-data-devices\") pod \"31ef19db-14de-48b2-b960-5c2dbffbe85d\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.797911 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31ef19db-14de-48b2-b960-5c2dbffbe85d-etc-swift\") pod \"31ef19db-14de-48b2-b960-5c2dbffbe85d\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.798031 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-scripts\") pod \"31ef19db-14de-48b2-b960-5c2dbffbe85d\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.798197 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-dispersionconf\") pod \"31ef19db-14de-48b2-b960-5c2dbffbe85d\" (UID: \"31ef19db-14de-48b2-b960-5c2dbffbe85d\") " Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.798538 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "31ef19db-14de-48b2-b960-5c2dbffbe85d" (UID: "31ef19db-14de-48b2-b960-5c2dbffbe85d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.798730 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.799041 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31ef19db-14de-48b2-b960-5c2dbffbe85d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "31ef19db-14de-48b2-b960-5c2dbffbe85d" (UID: "31ef19db-14de-48b2-b960-5c2dbffbe85d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.803967 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31ef19db-14de-48b2-b960-5c2dbffbe85d-kube-api-access-6hgtg" (OuterVolumeSpecName: "kube-api-access-6hgtg") pod "31ef19db-14de-48b2-b960-5c2dbffbe85d" (UID: "31ef19db-14de-48b2-b960-5c2dbffbe85d"). InnerVolumeSpecName "kube-api-access-6hgtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.813256 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.821797 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-scripts" (OuterVolumeSpecName: "scripts") pod "31ef19db-14de-48b2-b960-5c2dbffbe85d" (UID: "31ef19db-14de-48b2-b960-5c2dbffbe85d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.826000 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "31ef19db-14de-48b2-b960-5c2dbffbe85d" (UID: "31ef19db-14de-48b2-b960-5c2dbffbe85d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.826218 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "31ef19db-14de-48b2-b960-5c2dbffbe85d" (UID: "31ef19db-14de-48b2-b960-5c2dbffbe85d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.899953 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hgtg\" (UniqueName: \"kubernetes.io/projected/31ef19db-14de-48b2-b960-5c2dbffbe85d-kube-api-access-6hgtg\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.899993 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.900003 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31ef19db-14de-48b2-b960-5c2dbffbe85d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.900012 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31ef19db-14de-48b2-b960-5c2dbffbe85d-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:37 crc kubenswrapper[4692]: I0309 09:55:37.900022 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31ef19db-14de-48b2-b960-5c2dbffbe85d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.001241 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86wc8\" (UniqueName: \"kubernetes.io/projected/83338b51-8ebd-45e5-bee1-d20b036c37fc-kube-api-access-86wc8\") pod \"83338b51-8ebd-45e5-bee1-d20b036c37fc\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.001963 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-utilities\") pod \"83338b51-8ebd-45e5-bee1-d20b036c37fc\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.001991 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-catalog-content\") pod \"83338b51-8ebd-45e5-bee1-d20b036c37fc\" (UID: \"83338b51-8ebd-45e5-bee1-d20b036c37fc\") " Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.002689 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-utilities" (OuterVolumeSpecName: "utilities") pod "83338b51-8ebd-45e5-bee1-d20b036c37fc" (UID: "83338b51-8ebd-45e5-bee1-d20b036c37fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.049961 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83338b51-8ebd-45e5-bee1-d20b036c37fc" (UID: "83338b51-8ebd-45e5-bee1-d20b036c37fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.080593 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31ef19db-14de-48b2-b960-5c2dbffbe85d" path="/var/lib/kubelet/pods/31ef19db-14de-48b2-b960-5c2dbffbe85d/volumes" Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.103322 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.103548 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83338b51-8ebd-45e5-bee1-d20b036c37fc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.323603 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4x6md" Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.326583 4692 generic.go:334] "Generic (PLEG): container finished" podID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerID="2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51" exitCode=0 Mar 09 09:55:38 crc kubenswrapper[4692]: I0309 09:55:38.326738 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-28tcl" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.251365 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83338b51-8ebd-45e5-bee1-d20b036c37fc-kube-api-access-86wc8" (OuterVolumeSpecName: "kube-api-access-86wc8") pod "83338b51-8ebd-45e5-bee1-d20b036c37fc" (UID: "83338b51-8ebd-45e5-bee1-d20b036c37fc"). InnerVolumeSpecName "kube-api-access-86wc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.338970 4692 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.268s" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.341612 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86wc8\" (UniqueName: \"kubernetes.io/projected/83338b51-8ebd-45e5-bee1-d20b036c37fc-kube-api-access-86wc8\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.350308 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28tcl" event={"ID":"83338b51-8ebd-45e5-bee1-d20b036c37fc","Type":"ContainerDied","Data":"2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51"} Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.350608 4692 scope.go:117] "RemoveContainer" containerID="5fe7772552024d4bf9decec7c373bee2a1eecc3ab2ae65b2d1687934a7109b34" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.350888 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-28tcl" event={"ID":"83338b51-8ebd-45e5-bee1-d20b036c37fc","Type":"ContainerDied","Data":"2dd4ae4a4ee4b8cbde7ab762353c9e3c45b405f297b1ac2b9140e125ef8b4a3c"} Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.350921 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb"] Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.351229 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="registry-server" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.351249 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="registry-server" Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.351322 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="extract-content" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.351350 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="extract-content" Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.351399 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31ef19db-14de-48b2-b960-5c2dbffbe85d" containerName="swift-ring-rebalance" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.351410 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="31ef19db-14de-48b2-b960-5c2dbffbe85d" containerName="swift-ring-rebalance" Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.351426 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="extract-utilities" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.351433 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="extract-utilities" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.351606 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" containerName="registry-server" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.351662 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="31ef19db-14de-48b2-b960-5c2dbffbe85d" containerName="swift-ring-rebalance" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.352632 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb"] Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.352706 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.355304 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.355405 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.406178 4692 scope.go:117] "RemoveContainer" containerID="2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.425876 4692 scope.go:117] "RemoveContainer" containerID="7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.444027 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-28tcl"] Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.449782 4692 scope.go:117] "RemoveContainer" containerID="65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.450125 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-28tcl"] Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.492462 4692 scope.go:117] "RemoveContainer" containerID="2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51" Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.493030 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51\": container with ID starting with 2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51 not found: ID does not exist" containerID="2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.493061 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51"} err="failed to get container status \"2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51\": rpc error: code = NotFound desc = could not find container \"2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51\": container with ID starting with 2cac19e7d6a531b435f4875fa75d4935ef9d9420494b61b31d9c06fbb321ad51 not found: ID does not exist" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.493081 4692 scope.go:117] "RemoveContainer" containerID="7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572" Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.493321 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572\": container with ID starting with 7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572 not found: ID does not exist" containerID="7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.493345 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572"} err="failed to get container status \"7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572\": rpc error: code = NotFound desc = could not find container \"7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572\": container with ID starting with 7124248b99436e6563ea4e973a57e3e22430f7dcbbe6747dcd063f090df4f572 not found: ID does not exist" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.493360 4692 scope.go:117] "RemoveContainer" containerID="65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2" Mar 09 09:55:40 crc kubenswrapper[4692]: E0309 09:55:40.493659 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2\": container with ID starting with 65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2 not found: ID does not exist" containerID="65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.493713 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2"} err="failed to get container status \"65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2\": rpc error: code = NotFound desc = could not find container \"65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2\": container with ID starting with 65560fad39a145e7193467dddade6585ff5cbadb8da94af2b07a930d82d990b2 not found: ID does not exist" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.545997 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-swiftconf\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.546058 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-dispersionconf\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.546512 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49bbc\" (UniqueName: \"kubernetes.io/projected/47e4d47b-8d07-4db0-8b3b-d6649e290748-kube-api-access-49bbc\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.546592 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-scripts\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.546716 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47e4d47b-8d07-4db0-8b3b-d6649e290748-etc-swift\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.546761 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-ring-data-devices\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.647740 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-swiftconf\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.647809 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-dispersionconf\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.647849 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49bbc\" (UniqueName: \"kubernetes.io/projected/47e4d47b-8d07-4db0-8b3b-d6649e290748-kube-api-access-49bbc\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.647874 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-scripts\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.647912 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47e4d47b-8d07-4db0-8b3b-d6649e290748-etc-swift\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.647942 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-ring-data-devices\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.648760 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-ring-data-devices\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.649443 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47e4d47b-8d07-4db0-8b3b-d6649e290748-etc-swift\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.649450 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-scripts\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.653346 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-swiftconf\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.660992 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-dispersionconf\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.664389 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49bbc\" (UniqueName: \"kubernetes.io/projected/47e4d47b-8d07-4db0-8b3b-d6649e290748-kube-api-access-49bbc\") pod \"swift-ring-rebalance-debug-qhbwb\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:40 crc kubenswrapper[4692]: I0309 09:55:40.708335 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:41 crc kubenswrapper[4692]: I0309 09:55:41.117994 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb"] Mar 09 09:55:41 crc kubenswrapper[4692]: I0309 09:55:41.361039 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" event={"ID":"47e4d47b-8d07-4db0-8b3b-d6649e290748","Type":"ContainerStarted","Data":"31882b07d7ba81dbdf1376ed84debb308679d145ebcc592ffb10507fd63ac4a4"} Mar 09 09:55:41 crc kubenswrapper[4692]: I0309 09:55:41.361084 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" event={"ID":"47e4d47b-8d07-4db0-8b3b-d6649e290748","Type":"ContainerStarted","Data":"9bb1c6ce2c4bc85bdb4f04d0ca6e5aa6fdaafebc87d5f90f89db93cd78f96633"} Mar 09 09:55:41 crc kubenswrapper[4692]: I0309 09:55:41.381528 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" podStartSLOduration=3.381510963 podStartE2EDuration="3.381510963s" podCreationTimestamp="2026-03-09 09:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:41.378081256 +0000 UTC m=+2142.202816857" watchObservedRunningTime="2026-03-09 09:55:41.381510963 +0000 UTC m=+2142.206246564" Mar 09 09:55:42 crc kubenswrapper[4692]: I0309 09:55:42.081545 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83338b51-8ebd-45e5-bee1-d20b036c37fc" path="/var/lib/kubelet/pods/83338b51-8ebd-45e5-bee1-d20b036c37fc/volumes" Mar 09 09:55:43 crc kubenswrapper[4692]: I0309 09:55:43.379344 4692 generic.go:334] "Generic (PLEG): container finished" podID="47e4d47b-8d07-4db0-8b3b-d6649e290748" containerID="31882b07d7ba81dbdf1376ed84debb308679d145ebcc592ffb10507fd63ac4a4" exitCode=0 Mar 09 09:55:43 crc kubenswrapper[4692]: I0309 09:55:43.379386 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" event={"ID":"47e4d47b-8d07-4db0-8b3b-d6649e290748","Type":"ContainerDied","Data":"31882b07d7ba81dbdf1376ed84debb308679d145ebcc592ffb10507fd63ac4a4"} Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.699664 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.735231 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb"] Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.742444 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb"] Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.841924 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49bbc\" (UniqueName: \"kubernetes.io/projected/47e4d47b-8d07-4db0-8b3b-d6649e290748-kube-api-access-49bbc\") pod \"47e4d47b-8d07-4db0-8b3b-d6649e290748\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.842296 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47e4d47b-8d07-4db0-8b3b-d6649e290748-etc-swift\") pod \"47e4d47b-8d07-4db0-8b3b-d6649e290748\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.842487 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-scripts\") pod \"47e4d47b-8d07-4db0-8b3b-d6649e290748\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.842529 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-swiftconf\") pod \"47e4d47b-8d07-4db0-8b3b-d6649e290748\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.842563 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-dispersionconf\") pod \"47e4d47b-8d07-4db0-8b3b-d6649e290748\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.842601 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-ring-data-devices\") pod \"47e4d47b-8d07-4db0-8b3b-d6649e290748\" (UID: \"47e4d47b-8d07-4db0-8b3b-d6649e290748\") " Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.843347 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "47e4d47b-8d07-4db0-8b3b-d6649e290748" (UID: "47e4d47b-8d07-4db0-8b3b-d6649e290748"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.844593 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47e4d47b-8d07-4db0-8b3b-d6649e290748-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "47e4d47b-8d07-4db0-8b3b-d6649e290748" (UID: "47e4d47b-8d07-4db0-8b3b-d6649e290748"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.850139 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47e4d47b-8d07-4db0-8b3b-d6649e290748-kube-api-access-49bbc" (OuterVolumeSpecName: "kube-api-access-49bbc") pod "47e4d47b-8d07-4db0-8b3b-d6649e290748" (UID: "47e4d47b-8d07-4db0-8b3b-d6649e290748"). InnerVolumeSpecName "kube-api-access-49bbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.864939 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-scripts" (OuterVolumeSpecName: "scripts") pod "47e4d47b-8d07-4db0-8b3b-d6649e290748" (UID: "47e4d47b-8d07-4db0-8b3b-d6649e290748"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.870349 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "47e4d47b-8d07-4db0-8b3b-d6649e290748" (UID: "47e4d47b-8d07-4db0-8b3b-d6649e290748"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.873204 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "47e4d47b-8d07-4db0-8b3b-d6649e290748" (UID: "47e4d47b-8d07-4db0-8b3b-d6649e290748"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.944235 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.944291 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/47e4d47b-8d07-4db0-8b3b-d6649e290748-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.944303 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.944315 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49bbc\" (UniqueName: \"kubernetes.io/projected/47e4d47b-8d07-4db0-8b3b-d6649e290748-kube-api-access-49bbc\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.944348 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/47e4d47b-8d07-4db0-8b3b-d6649e290748-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:44 crc kubenswrapper[4692]: I0309 09:55:44.944361 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/47e4d47b-8d07-4db0-8b3b-d6649e290748-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.397854 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bb1c6ce2c4bc85bdb4f04d0ca6e5aa6fdaafebc87d5f90f89db93cd78f96633" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.397927 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qhbwb" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.889258 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5"] Mar 09 09:55:45 crc kubenswrapper[4692]: E0309 09:55:45.889691 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47e4d47b-8d07-4db0-8b3b-d6649e290748" containerName="swift-ring-rebalance" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.889704 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="47e4d47b-8d07-4db0-8b3b-d6649e290748" containerName="swift-ring-rebalance" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.889837 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="47e4d47b-8d07-4db0-8b3b-d6649e290748" containerName="swift-ring-rebalance" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.890588 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.898387 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.901026 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5"] Mar 09 09:55:45 crc kubenswrapper[4692]: I0309 09:55:45.901305 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.059189 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-dispersionconf\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.059660 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-swiftconf\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.059690 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-ring-data-devices\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.059720 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2pxr\" (UniqueName: \"kubernetes.io/projected/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-kube-api-access-h2pxr\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.059771 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-scripts\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.059999 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-etc-swift\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.091140 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47e4d47b-8d07-4db0-8b3b-d6649e290748" path="/var/lib/kubelet/pods/47e4d47b-8d07-4db0-8b3b-d6649e290748/volumes" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.161132 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-swiftconf\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.161208 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-ring-data-devices\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.161243 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2pxr\" (UniqueName: \"kubernetes.io/projected/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-kube-api-access-h2pxr\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.161307 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-scripts\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.161377 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-etc-swift\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.161419 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-dispersionconf\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.162265 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-ring-data-devices\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.162359 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-etc-swift\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.162612 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-scripts\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.167405 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-dispersionconf\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.167404 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-swiftconf\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.181473 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2pxr\" (UniqueName: \"kubernetes.io/projected/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-kube-api-access-h2pxr\") pod \"swift-ring-rebalance-debug-qjnm5\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.209498 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:46 crc kubenswrapper[4692]: I0309 09:55:46.631366 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5"] Mar 09 09:55:47 crc kubenswrapper[4692]: I0309 09:55:47.415262 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" event={"ID":"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd","Type":"ContainerStarted","Data":"d9c455e2eb472f994f09d5320bc43fac01669569ca4b4c8104176453d3b67ec5"} Mar 09 09:55:47 crc kubenswrapper[4692]: I0309 09:55:47.415315 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" event={"ID":"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd","Type":"ContainerStarted","Data":"6be7989569474a2784d33febf671e522b5ebaa09af302db0934f6ee326b29683"} Mar 09 09:55:47 crc kubenswrapper[4692]: I0309 09:55:47.434303 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" podStartSLOduration=2.434287902 podStartE2EDuration="2.434287902s" podCreationTimestamp="2026-03-09 09:55:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:47.433327545 +0000 UTC m=+2148.258063136" watchObservedRunningTime="2026-03-09 09:55:47.434287902 +0000 UTC m=+2148.259023483" Mar 09 09:55:48 crc kubenswrapper[4692]: I0309 09:55:48.424253 4692 generic.go:334] "Generic (PLEG): container finished" podID="ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" containerID="d9c455e2eb472f994f09d5320bc43fac01669569ca4b4c8104176453d3b67ec5" exitCode=0 Mar 09 09:55:48 crc kubenswrapper[4692]: I0309 09:55:48.424298 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" event={"ID":"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd","Type":"ContainerDied","Data":"d9c455e2eb472f994f09d5320bc43fac01669569ca4b4c8104176453d3b67ec5"} Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.717336 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.752558 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5"] Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.759885 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5"] Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.823996 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-scripts\") pod \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.824051 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2pxr\" (UniqueName: \"kubernetes.io/projected/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-kube-api-access-h2pxr\") pod \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.824080 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-swiftconf\") pod \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.824097 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-ring-data-devices\") pod \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.824188 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-dispersionconf\") pod \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.824221 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-etc-swift\") pod \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\" (UID: \"ef0b41c0-58b2-4599-8f5b-dbd5913c40dd\") " Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.825124 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" (UID: "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.825553 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" (UID: "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.831389 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-kube-api-access-h2pxr" (OuterVolumeSpecName: "kube-api-access-h2pxr") pod "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" (UID: "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd"). InnerVolumeSpecName "kube-api-access-h2pxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.849050 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-scripts" (OuterVolumeSpecName: "scripts") pod "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" (UID: "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.851290 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" (UID: "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.858345 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" (UID: "ef0b41c0-58b2-4599-8f5b-dbd5913c40dd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.925363 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.925690 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2pxr\" (UniqueName: \"kubernetes.io/projected/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-kube-api-access-h2pxr\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.925704 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.925716 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.925727 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:49 crc kubenswrapper[4692]: I0309 09:55:49.925736 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.081064 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" path="/var/lib/kubelet/pods/ef0b41c0-58b2-4599-8f5b-dbd5913c40dd/volumes" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.445745 4692 scope.go:117] "RemoveContainer" containerID="d9c455e2eb472f994f09d5320bc43fac01669569ca4b4c8104176453d3b67ec5" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.445860 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qjnm5" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.903628 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fct85"] Mar 09 09:55:50 crc kubenswrapper[4692]: E0309 09:55:50.903996 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" containerName="swift-ring-rebalance" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.904015 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" containerName="swift-ring-rebalance" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.904198 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef0b41c0-58b2-4599-8f5b-dbd5913c40dd" containerName="swift-ring-rebalance" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.904790 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.908096 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.908912 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.915744 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fct85"] Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.941242 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-swiftconf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.941315 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7e2ad4d5-965b-466e-8ae4-458998d4281e-etc-swift\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.941601 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-ring-data-devices\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.941633 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxznf\" (UniqueName: \"kubernetes.io/projected/7e2ad4d5-965b-466e-8ae4-458998d4281e-kube-api-access-qxznf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.941781 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-scripts\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:50 crc kubenswrapper[4692]: I0309 09:55:50.941818 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-dispersionconf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.042817 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-ring-data-devices\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.042870 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxznf\" (UniqueName: \"kubernetes.io/projected/7e2ad4d5-965b-466e-8ae4-458998d4281e-kube-api-access-qxznf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.042910 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-scripts\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.042936 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-dispersionconf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.042975 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-swiftconf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.043004 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7e2ad4d5-965b-466e-8ae4-458998d4281e-etc-swift\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.044065 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7e2ad4d5-965b-466e-8ae4-458998d4281e-etc-swift\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.044215 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-scripts\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.044215 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-ring-data-devices\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.048685 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-dispersionconf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.048685 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-swiftconf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.065566 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxznf\" (UniqueName: \"kubernetes.io/projected/7e2ad4d5-965b-466e-8ae4-458998d4281e-kube-api-access-qxznf\") pod \"swift-ring-rebalance-debug-fct85\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.226127 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:51 crc kubenswrapper[4692]: I0309 09:55:51.683939 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fct85"] Mar 09 09:55:52 crc kubenswrapper[4692]: I0309 09:55:52.467867 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" event={"ID":"7e2ad4d5-965b-466e-8ae4-458998d4281e","Type":"ContainerStarted","Data":"145a219aa7ba64540dc0eb3092045f87edb0172d48d2a2b4ad39834c066a3381"} Mar 09 09:55:52 crc kubenswrapper[4692]: I0309 09:55:52.468232 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" event={"ID":"7e2ad4d5-965b-466e-8ae4-458998d4281e","Type":"ContainerStarted","Data":"960da1325265f416aa2cd563471f3fce470d4d4b2871e0f0af1b559e089042d0"} Mar 09 09:55:53 crc kubenswrapper[4692]: I0309 09:55:53.478233 4692 generic.go:334] "Generic (PLEG): container finished" podID="7e2ad4d5-965b-466e-8ae4-458998d4281e" containerID="145a219aa7ba64540dc0eb3092045f87edb0172d48d2a2b4ad39834c066a3381" exitCode=0 Mar 09 09:55:53 crc kubenswrapper[4692]: I0309 09:55:53.478296 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" event={"ID":"7e2ad4d5-965b-466e-8ae4-458998d4281e","Type":"ContainerDied","Data":"145a219aa7ba64540dc0eb3092045f87edb0172d48d2a2b4ad39834c066a3381"} Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.812219 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.852802 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fct85"] Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.859144 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-fct85"] Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.913075 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-dispersionconf\") pod \"7e2ad4d5-965b-466e-8ae4-458998d4281e\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.913137 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxznf\" (UniqueName: \"kubernetes.io/projected/7e2ad4d5-965b-466e-8ae4-458998d4281e-kube-api-access-qxznf\") pod \"7e2ad4d5-965b-466e-8ae4-458998d4281e\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.913187 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7e2ad4d5-965b-466e-8ae4-458998d4281e-etc-swift\") pod \"7e2ad4d5-965b-466e-8ae4-458998d4281e\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.913210 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-swiftconf\") pod \"7e2ad4d5-965b-466e-8ae4-458998d4281e\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.913287 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-ring-data-devices\") pod \"7e2ad4d5-965b-466e-8ae4-458998d4281e\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.913326 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-scripts\") pod \"7e2ad4d5-965b-466e-8ae4-458998d4281e\" (UID: \"7e2ad4d5-965b-466e-8ae4-458998d4281e\") " Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.913967 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7e2ad4d5-965b-466e-8ae4-458998d4281e" (UID: "7e2ad4d5-965b-466e-8ae4-458998d4281e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.914227 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e2ad4d5-965b-466e-8ae4-458998d4281e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7e2ad4d5-965b-466e-8ae4-458998d4281e" (UID: "7e2ad4d5-965b-466e-8ae4-458998d4281e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.921656 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2ad4d5-965b-466e-8ae4-458998d4281e-kube-api-access-qxznf" (OuterVolumeSpecName: "kube-api-access-qxznf") pod "7e2ad4d5-965b-466e-8ae4-458998d4281e" (UID: "7e2ad4d5-965b-466e-8ae4-458998d4281e"). InnerVolumeSpecName "kube-api-access-qxznf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.937244 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-scripts" (OuterVolumeSpecName: "scripts") pod "7e2ad4d5-965b-466e-8ae4-458998d4281e" (UID: "7e2ad4d5-965b-466e-8ae4-458998d4281e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.940226 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7e2ad4d5-965b-466e-8ae4-458998d4281e" (UID: "7e2ad4d5-965b-466e-8ae4-458998d4281e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:54 crc kubenswrapper[4692]: I0309 09:55:54.940972 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7e2ad4d5-965b-466e-8ae4-458998d4281e" (UID: "7e2ad4d5-965b-466e-8ae4-458998d4281e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.015077 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.015137 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxznf\" (UniqueName: \"kubernetes.io/projected/7e2ad4d5-965b-466e-8ae4-458998d4281e-kube-api-access-qxznf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.015151 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7e2ad4d5-965b-466e-8ae4-458998d4281e-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.015176 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7e2ad4d5-965b-466e-8ae4-458998d4281e-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.015186 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.015196 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e2ad4d5-965b-466e-8ae4-458998d4281e-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.494758 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="960da1325265f416aa2cd563471f3fce470d4d4b2871e0f0af1b559e089042d0" Mar 09 09:55:55 crc kubenswrapper[4692]: I0309 09:55:55.494841 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-fct85" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.005231 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq"] Mar 09 09:55:56 crc kubenswrapper[4692]: E0309 09:55:56.005782 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2ad4d5-965b-466e-8ae4-458998d4281e" containerName="swift-ring-rebalance" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.005802 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2ad4d5-965b-466e-8ae4-458998d4281e" containerName="swift-ring-rebalance" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.006027 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2ad4d5-965b-466e-8ae4-458998d4281e" containerName="swift-ring-rebalance" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.006867 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.010536 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.010592 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.015032 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq"] Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.027995 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6ljz\" (UniqueName: \"kubernetes.io/projected/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-kube-api-access-s6ljz\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.028098 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-scripts\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.028145 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-swiftconf\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.028221 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-ring-data-devices\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.028251 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-etc-swift\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.028282 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-dispersionconf\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.080640 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e2ad4d5-965b-466e-8ae4-458998d4281e" path="/var/lib/kubelet/pods/7e2ad4d5-965b-466e-8ae4-458998d4281e/volumes" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.134426 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-scripts\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.134772 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-swiftconf\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.134973 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-ring-data-devices\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.135120 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-etc-swift\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.135258 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-dispersionconf\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.135427 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6ljz\" (UniqueName: \"kubernetes.io/projected/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-kube-api-access-s6ljz\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.136054 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-ring-data-devices\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.136643 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-scripts\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.137056 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-etc-swift\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.141543 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-dispersionconf\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.149088 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-swiftconf\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.163558 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6ljz\" (UniqueName: \"kubernetes.io/projected/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-kube-api-access-s6ljz\") pod \"swift-ring-rebalance-debug-ddzkq\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.327824 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:56 crc kubenswrapper[4692]: I0309 09:55:56.745419 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq"] Mar 09 09:55:56 crc kubenswrapper[4692]: W0309 09:55:56.747869 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f74d217_798a_4de0_9f0d_4fb11cbf76ca.slice/crio-356f569499e4d300cb450bc2cabc3c00c0fe1e92482389f3f920ed00a6e46d40 WatchSource:0}: Error finding container 356f569499e4d300cb450bc2cabc3c00c0fe1e92482389f3f920ed00a6e46d40: Status 404 returned error can't find the container with id 356f569499e4d300cb450bc2cabc3c00c0fe1e92482389f3f920ed00a6e46d40 Mar 09 09:55:57 crc kubenswrapper[4692]: I0309 09:55:57.511789 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" event={"ID":"2f74d217-798a-4de0-9f0d-4fb11cbf76ca","Type":"ContainerStarted","Data":"5744bcebae2a0bc3fba1b3e4ec08e10cad706c5c303e6bb0e30d776e6be36fc3"} Mar 09 09:55:57 crc kubenswrapper[4692]: I0309 09:55:57.513146 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" event={"ID":"2f74d217-798a-4de0-9f0d-4fb11cbf76ca","Type":"ContainerStarted","Data":"356f569499e4d300cb450bc2cabc3c00c0fe1e92482389f3f920ed00a6e46d40"} Mar 09 09:55:57 crc kubenswrapper[4692]: I0309 09:55:57.532995 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" podStartSLOduration=2.5329788840000003 podStartE2EDuration="2.532978884s" podCreationTimestamp="2026-03-09 09:55:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:55:57.529049173 +0000 UTC m=+2158.353784774" watchObservedRunningTime="2026-03-09 09:55:57.532978884 +0000 UTC m=+2158.357714465" Mar 09 09:55:58 crc kubenswrapper[4692]: I0309 09:55:58.524781 4692 generic.go:334] "Generic (PLEG): container finished" podID="2f74d217-798a-4de0-9f0d-4fb11cbf76ca" containerID="5744bcebae2a0bc3fba1b3e4ec08e10cad706c5c303e6bb0e30d776e6be36fc3" exitCode=0 Mar 09 09:55:58 crc kubenswrapper[4692]: I0309 09:55:58.524981 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" event={"ID":"2f74d217-798a-4de0-9f0d-4fb11cbf76ca","Type":"ContainerDied","Data":"5744bcebae2a0bc3fba1b3e4ec08e10cad706c5c303e6bb0e30d776e6be36fc3"} Mar 09 09:55:59 crc kubenswrapper[4692]: I0309 09:55:59.863593 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:55:59 crc kubenswrapper[4692]: I0309 09:55:59.896226 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq"] Mar 09 09:55:59 crc kubenswrapper[4692]: I0309 09:55:59.900490 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq"] Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.019325 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6ljz\" (UniqueName: \"kubernetes.io/projected/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-kube-api-access-s6ljz\") pod \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.019409 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-etc-swift\") pod \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.019482 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-dispersionconf\") pod \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.019581 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-swiftconf\") pod \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.019619 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-scripts\") pod \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.019815 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-ring-data-devices\") pod \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\" (UID: \"2f74d217-798a-4de0-9f0d-4fb11cbf76ca\") " Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.020625 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2f74d217-798a-4de0-9f0d-4fb11cbf76ca" (UID: "2f74d217-798a-4de0-9f0d-4fb11cbf76ca"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.021003 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2f74d217-798a-4de0-9f0d-4fb11cbf76ca" (UID: "2f74d217-798a-4de0-9f0d-4fb11cbf76ca"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.025971 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-kube-api-access-s6ljz" (OuterVolumeSpecName: "kube-api-access-s6ljz") pod "2f74d217-798a-4de0-9f0d-4fb11cbf76ca" (UID: "2f74d217-798a-4de0-9f0d-4fb11cbf76ca"). InnerVolumeSpecName "kube-api-access-s6ljz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.042874 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-scripts" (OuterVolumeSpecName: "scripts") pod "2f74d217-798a-4de0-9f0d-4fb11cbf76ca" (UID: "2f74d217-798a-4de0-9f0d-4fb11cbf76ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.042985 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2f74d217-798a-4de0-9f0d-4fb11cbf76ca" (UID: "2f74d217-798a-4de0-9f0d-4fb11cbf76ca"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.045034 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2f74d217-798a-4de0-9f0d-4fb11cbf76ca" (UID: "2f74d217-798a-4de0-9f0d-4fb11cbf76ca"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.095037 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f74d217-798a-4de0-9f0d-4fb11cbf76ca" path="/var/lib/kubelet/pods/2f74d217-798a-4de0-9f0d-4fb11cbf76ca/volumes" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.125319 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.125359 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.125372 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.125386 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6ljz\" (UniqueName: \"kubernetes.io/projected/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-kube-api-access-s6ljz\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.125396 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.125405 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f74d217-798a-4de0-9f0d-4fb11cbf76ca-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.140444 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550836-glwp9"] Mar 09 09:56:00 crc kubenswrapper[4692]: E0309 09:56:00.142863 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f74d217-798a-4de0-9f0d-4fb11cbf76ca" containerName="swift-ring-rebalance" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.143495 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f74d217-798a-4de0-9f0d-4fb11cbf76ca" containerName="swift-ring-rebalance" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.143871 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f74d217-798a-4de0-9f0d-4fb11cbf76ca" containerName="swift-ring-rebalance" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.144543 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550836-glwp9" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.147096 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.147226 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550836-glwp9"] Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.147383 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.148290 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.327853 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgrd9\" (UniqueName: \"kubernetes.io/projected/a6d3a680-5cd9-487b-a928-b83c4f24728c-kube-api-access-hgrd9\") pod \"auto-csr-approver-29550836-glwp9\" (UID: \"a6d3a680-5cd9-487b-a928-b83c4f24728c\") " pod="openshift-infra/auto-csr-approver-29550836-glwp9" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.429503 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgrd9\" (UniqueName: \"kubernetes.io/projected/a6d3a680-5cd9-487b-a928-b83c4f24728c-kube-api-access-hgrd9\") pod \"auto-csr-approver-29550836-glwp9\" (UID: \"a6d3a680-5cd9-487b-a928-b83c4f24728c\") " pod="openshift-infra/auto-csr-approver-29550836-glwp9" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.449100 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgrd9\" (UniqueName: \"kubernetes.io/projected/a6d3a680-5cd9-487b-a928-b83c4f24728c-kube-api-access-hgrd9\") pod \"auto-csr-approver-29550836-glwp9\" (UID: \"a6d3a680-5cd9-487b-a928-b83c4f24728c\") " pod="openshift-infra/auto-csr-approver-29550836-glwp9" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.467809 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550836-glwp9" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.546570 4692 scope.go:117] "RemoveContainer" containerID="5744bcebae2a0bc3fba1b3e4ec08e10cad706c5c303e6bb0e30d776e6be36fc3" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.546695 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ddzkq" Mar 09 09:56:00 crc kubenswrapper[4692]: I0309 09:56:00.881317 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550836-glwp9"] Mar 09 09:56:00 crc kubenswrapper[4692]: W0309 09:56:00.886442 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6d3a680_5cd9_487b_a928_b83c4f24728c.slice/crio-facd44843381db378482b2697a28a3278eeb0e720f56029b6280cedd2c21380d WatchSource:0}: Error finding container facd44843381db378482b2697a28a3278eeb0e720f56029b6280cedd2c21380d: Status 404 returned error can't find the container with id facd44843381db378482b2697a28a3278eeb0e720f56029b6280cedd2c21380d Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.030213 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b"] Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.031360 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.033809 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.033871 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.040493 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-swiftconf\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.040543 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qvs4\" (UniqueName: \"kubernetes.io/projected/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-kube-api-access-8qvs4\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.040583 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-ring-data-devices\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.040609 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-etc-swift\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.040646 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-dispersionconf\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.040683 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-scripts\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.041922 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b"] Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.141689 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-swiftconf\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.141737 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qvs4\" (UniqueName: \"kubernetes.io/projected/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-kube-api-access-8qvs4\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.141767 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-ring-data-devices\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.141794 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-etc-swift\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.141846 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-dispersionconf\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.141883 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-scripts\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.142702 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-scripts\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.143658 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-ring-data-devices\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.143671 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-etc-swift\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.148897 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-swiftconf\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.150571 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-dispersionconf\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.159325 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qvs4\" (UniqueName: \"kubernetes.io/projected/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-kube-api-access-8qvs4\") pod \"swift-ring-rebalance-debug-x6p2b\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.358639 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.567418 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550836-glwp9" event={"ID":"a6d3a680-5cd9-487b-a928-b83c4f24728c","Type":"ContainerStarted","Data":"facd44843381db378482b2697a28a3278eeb0e720f56029b6280cedd2c21380d"} Mar 09 09:56:01 crc kubenswrapper[4692]: I0309 09:56:01.773618 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b"] Mar 09 09:56:02 crc kubenswrapper[4692]: I0309 09:56:02.583310 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" event={"ID":"1d70d6f7-dea3-4a12-b081-0f0ed49346d0","Type":"ContainerStarted","Data":"a0d05f3c2440ecd9349096613efdff0e763516fc23ce41736cde0d954964667c"} Mar 09 09:56:02 crc kubenswrapper[4692]: I0309 09:56:02.583627 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" event={"ID":"1d70d6f7-dea3-4a12-b081-0f0ed49346d0","Type":"ContainerStarted","Data":"48e25ca32c26675f9c448781b954eebc72ca8d2411d6d17749f8e07e47b6ed9a"} Mar 09 09:56:02 crc kubenswrapper[4692]: I0309 09:56:02.587123 4692 generic.go:334] "Generic (PLEG): container finished" podID="a6d3a680-5cd9-487b-a928-b83c4f24728c" containerID="d1858c5b055955c720712dbcb9b715bd829de6079b9294b70e909c284eca4085" exitCode=0 Mar 09 09:56:02 crc kubenswrapper[4692]: I0309 09:56:02.587203 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550836-glwp9" event={"ID":"a6d3a680-5cd9-487b-a928-b83c4f24728c","Type":"ContainerDied","Data":"d1858c5b055955c720712dbcb9b715bd829de6079b9294b70e909c284eca4085"} Mar 09 09:56:02 crc kubenswrapper[4692]: I0309 09:56:02.608514 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" podStartSLOduration=1.608484751 podStartE2EDuration="1.608484751s" podCreationTimestamp="2026-03-09 09:56:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:02.607123603 +0000 UTC m=+2163.431859184" watchObservedRunningTime="2026-03-09 09:56:02.608484751 +0000 UTC m=+2163.433220342" Mar 09 09:56:03 crc kubenswrapper[4692]: I0309 09:56:03.601265 4692 generic.go:334] "Generic (PLEG): container finished" podID="1d70d6f7-dea3-4a12-b081-0f0ed49346d0" containerID="a0d05f3c2440ecd9349096613efdff0e763516fc23ce41736cde0d954964667c" exitCode=0 Mar 09 09:56:03 crc kubenswrapper[4692]: I0309 09:56:03.601587 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" event={"ID":"1d70d6f7-dea3-4a12-b081-0f0ed49346d0","Type":"ContainerDied","Data":"a0d05f3c2440ecd9349096613efdff0e763516fc23ce41736cde0d954964667c"} Mar 09 09:56:03 crc kubenswrapper[4692]: I0309 09:56:03.954597 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550836-glwp9" Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.103533 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgrd9\" (UniqueName: \"kubernetes.io/projected/a6d3a680-5cd9-487b-a928-b83c4f24728c-kube-api-access-hgrd9\") pod \"a6d3a680-5cd9-487b-a928-b83c4f24728c\" (UID: \"a6d3a680-5cd9-487b-a928-b83c4f24728c\") " Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.109863 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6d3a680-5cd9-487b-a928-b83c4f24728c-kube-api-access-hgrd9" (OuterVolumeSpecName: "kube-api-access-hgrd9") pod "a6d3a680-5cd9-487b-a928-b83c4f24728c" (UID: "a6d3a680-5cd9-487b-a928-b83c4f24728c"). InnerVolumeSpecName "kube-api-access-hgrd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.206333 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgrd9\" (UniqueName: \"kubernetes.io/projected/a6d3a680-5cd9-487b-a928-b83c4f24728c-kube-api-access-hgrd9\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.613441 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550836-glwp9" Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.613426 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550836-glwp9" event={"ID":"a6d3a680-5cd9-487b-a928-b83c4f24728c","Type":"ContainerDied","Data":"facd44843381db378482b2697a28a3278eeb0e720f56029b6280cedd2c21380d"} Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.613589 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="facd44843381db378482b2697a28a3278eeb0e720f56029b6280cedd2c21380d" Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.899758 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.927749 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b"] Mar 09 09:56:04 crc kubenswrapper[4692]: I0309 09:56:04.934214 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b"] Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.015667 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550830-n66j7"] Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.017815 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-etc-swift\") pod \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.017864 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-dispersionconf\") pod \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.017929 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qvs4\" (UniqueName: \"kubernetes.io/projected/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-kube-api-access-8qvs4\") pod \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.017955 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-swiftconf\") pod \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.017996 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-ring-data-devices\") pod \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.018040 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-scripts\") pod \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\" (UID: \"1d70d6f7-dea3-4a12-b081-0f0ed49346d0\") " Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.019740 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1d70d6f7-dea3-4a12-b081-0f0ed49346d0" (UID: "1d70d6f7-dea3-4a12-b081-0f0ed49346d0"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.019837 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1d70d6f7-dea3-4a12-b081-0f0ed49346d0" (UID: "1d70d6f7-dea3-4a12-b081-0f0ed49346d0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.021864 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550830-n66j7"] Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.023486 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-kube-api-access-8qvs4" (OuterVolumeSpecName: "kube-api-access-8qvs4") pod "1d70d6f7-dea3-4a12-b081-0f0ed49346d0" (UID: "1d70d6f7-dea3-4a12-b081-0f0ed49346d0"). InnerVolumeSpecName "kube-api-access-8qvs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.041660 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-scripts" (OuterVolumeSpecName: "scripts") pod "1d70d6f7-dea3-4a12-b081-0f0ed49346d0" (UID: "1d70d6f7-dea3-4a12-b081-0f0ed49346d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.043561 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1d70d6f7-dea3-4a12-b081-0f0ed49346d0" (UID: "1d70d6f7-dea3-4a12-b081-0f0ed49346d0"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.048381 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1d70d6f7-dea3-4a12-b081-0f0ed49346d0" (UID: "1d70d6f7-dea3-4a12-b081-0f0ed49346d0"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.120243 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.120302 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.120318 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qvs4\" (UniqueName: \"kubernetes.io/projected/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-kube-api-access-8qvs4\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.120335 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.120349 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.120365 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d70d6f7-dea3-4a12-b081-0f0ed49346d0-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.626465 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48e25ca32c26675f9c448781b954eebc72ca8d2411d6d17749f8e07e47b6ed9a" Mar 09 09:56:05 crc kubenswrapper[4692]: I0309 09:56:05.626555 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-x6p2b" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.082267 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d70d6f7-dea3-4a12-b081-0f0ed49346d0" path="/var/lib/kubelet/pods/1d70d6f7-dea3-4a12-b081-0f0ed49346d0/volumes" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.082957 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bac7da7-e0f6-4875-b683-9dd6842c915b" path="/var/lib/kubelet/pods/4bac7da7-e0f6-4875-b683-9dd6842c915b/volumes" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.100655 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr"] Mar 09 09:56:06 crc kubenswrapper[4692]: E0309 09:56:06.100967 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d70d6f7-dea3-4a12-b081-0f0ed49346d0" containerName="swift-ring-rebalance" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.100989 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d70d6f7-dea3-4a12-b081-0f0ed49346d0" containerName="swift-ring-rebalance" Mar 09 09:56:06 crc kubenswrapper[4692]: E0309 09:56:06.101024 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6d3a680-5cd9-487b-a928-b83c4f24728c" containerName="oc" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.101033 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6d3a680-5cd9-487b-a928-b83c4f24728c" containerName="oc" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.101247 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6d3a680-5cd9-487b-a928-b83c4f24728c" containerName="oc" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.101274 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d70d6f7-dea3-4a12-b081-0f0ed49346d0" containerName="swift-ring-rebalance" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.101816 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.104234 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.104664 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.111661 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr"] Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.239380 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-swiftconf\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.239725 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq47w\" (UniqueName: \"kubernetes.io/projected/99a6cea1-df62-410b-ab13-7c64f2215633-kube-api-access-nq47w\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.239844 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-dispersionconf\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.239994 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-ring-data-devices\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.240104 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-scripts\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.240245 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99a6cea1-df62-410b-ab13-7c64f2215633-etc-swift\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.342294 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-ring-data-devices\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.342368 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-scripts\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.342426 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99a6cea1-df62-410b-ab13-7c64f2215633-etc-swift\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.342476 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-swiftconf\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.342499 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq47w\" (UniqueName: \"kubernetes.io/projected/99a6cea1-df62-410b-ab13-7c64f2215633-kube-api-access-nq47w\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.342559 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-dispersionconf\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.343557 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99a6cea1-df62-410b-ab13-7c64f2215633-etc-swift\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.343873 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-scripts\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.343898 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-ring-data-devices\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.349244 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-dispersionconf\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.349270 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-swiftconf\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.363938 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq47w\" (UniqueName: \"kubernetes.io/projected/99a6cea1-df62-410b-ab13-7c64f2215633-kube-api-access-nq47w\") pod \"swift-ring-rebalance-debug-jwlhr\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.427811 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:06 crc kubenswrapper[4692]: I0309 09:56:06.880572 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr"] Mar 09 09:56:06 crc kubenswrapper[4692]: W0309 09:56:06.889121 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a6cea1_df62_410b_ab13_7c64f2215633.slice/crio-54c20e0b9b8b52f431164912fbbdde3adca35c7f528fc287ce623fd4398a9779 WatchSource:0}: Error finding container 54c20e0b9b8b52f431164912fbbdde3adca35c7f528fc287ce623fd4398a9779: Status 404 returned error can't find the container with id 54c20e0b9b8b52f431164912fbbdde3adca35c7f528fc287ce623fd4398a9779 Mar 09 09:56:07 crc kubenswrapper[4692]: I0309 09:56:07.649007 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" event={"ID":"99a6cea1-df62-410b-ab13-7c64f2215633","Type":"ContainerStarted","Data":"70cfe070305c0f4e39019add4cc80e2d984641198f3650df24b01d3bf445116d"} Mar 09 09:56:07 crc kubenswrapper[4692]: I0309 09:56:07.649455 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" event={"ID":"99a6cea1-df62-410b-ab13-7c64f2215633","Type":"ContainerStarted","Data":"54c20e0b9b8b52f431164912fbbdde3adca35c7f528fc287ce623fd4398a9779"} Mar 09 09:56:07 crc kubenswrapper[4692]: I0309 09:56:07.673213 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" podStartSLOduration=1.6731885229999999 podStartE2EDuration="1.673188523s" podCreationTimestamp="2026-03-09 09:56:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:07.669696054 +0000 UTC m=+2168.494431635" watchObservedRunningTime="2026-03-09 09:56:07.673188523 +0000 UTC m=+2168.497924114" Mar 09 09:56:08 crc kubenswrapper[4692]: I0309 09:56:08.658572 4692 generic.go:334] "Generic (PLEG): container finished" podID="99a6cea1-df62-410b-ab13-7c64f2215633" containerID="70cfe070305c0f4e39019add4cc80e2d984641198f3650df24b01d3bf445116d" exitCode=0 Mar 09 09:56:08 crc kubenswrapper[4692]: I0309 09:56:08.658638 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" event={"ID":"99a6cea1-df62-410b-ab13-7c64f2215633","Type":"ContainerDied","Data":"70cfe070305c0f4e39019add4cc80e2d984641198f3650df24b01d3bf445116d"} Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.015204 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.040263 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99a6cea1-df62-410b-ab13-7c64f2215633-etc-swift\") pod \"99a6cea1-df62-410b-ab13-7c64f2215633\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.040345 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-dispersionconf\") pod \"99a6cea1-df62-410b-ab13-7c64f2215633\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.040453 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-scripts\") pod \"99a6cea1-df62-410b-ab13-7c64f2215633\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.040485 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-ring-data-devices\") pod \"99a6cea1-df62-410b-ab13-7c64f2215633\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.040556 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-swiftconf\") pod \"99a6cea1-df62-410b-ab13-7c64f2215633\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.040603 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq47w\" (UniqueName: \"kubernetes.io/projected/99a6cea1-df62-410b-ab13-7c64f2215633-kube-api-access-nq47w\") pod \"99a6cea1-df62-410b-ab13-7c64f2215633\" (UID: \"99a6cea1-df62-410b-ab13-7c64f2215633\") " Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.041391 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "99a6cea1-df62-410b-ab13-7c64f2215633" (UID: "99a6cea1-df62-410b-ab13-7c64f2215633"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.041831 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99a6cea1-df62-410b-ab13-7c64f2215633-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "99a6cea1-df62-410b-ab13-7c64f2215633" (UID: "99a6cea1-df62-410b-ab13-7c64f2215633"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.049722 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99a6cea1-df62-410b-ab13-7c64f2215633-kube-api-access-nq47w" (OuterVolumeSpecName: "kube-api-access-nq47w") pod "99a6cea1-df62-410b-ab13-7c64f2215633" (UID: "99a6cea1-df62-410b-ab13-7c64f2215633"). InnerVolumeSpecName "kube-api-access-nq47w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.055456 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr"] Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.063526 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr"] Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.067384 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-scripts" (OuterVolumeSpecName: "scripts") pod "99a6cea1-df62-410b-ab13-7c64f2215633" (UID: "99a6cea1-df62-410b-ab13-7c64f2215633"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.070699 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "99a6cea1-df62-410b-ab13-7c64f2215633" (UID: "99a6cea1-df62-410b-ab13-7c64f2215633"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.074652 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "99a6cea1-df62-410b-ab13-7c64f2215633" (UID: "99a6cea1-df62-410b-ab13-7c64f2215633"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.082206 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99a6cea1-df62-410b-ab13-7c64f2215633" path="/var/lib/kubelet/pods/99a6cea1-df62-410b-ab13-7c64f2215633/volumes" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.143635 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq47w\" (UniqueName: \"kubernetes.io/projected/99a6cea1-df62-410b-ab13-7c64f2215633-kube-api-access-nq47w\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.143704 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99a6cea1-df62-410b-ab13-7c64f2215633-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.143720 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.143733 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.143747 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99a6cea1-df62-410b-ab13-7c64f2215633-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.143761 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99a6cea1-df62-410b-ab13-7c64f2215633-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.680834 4692 scope.go:117] "RemoveContainer" containerID="70cfe070305c0f4e39019add4cc80e2d984641198f3650df24b01d3bf445116d" Mar 09 09:56:10 crc kubenswrapper[4692]: I0309 09:56:10.680904 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jwlhr" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.205941 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l7nff"] Mar 09 09:56:11 crc kubenswrapper[4692]: E0309 09:56:11.206761 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99a6cea1-df62-410b-ab13-7c64f2215633" containerName="swift-ring-rebalance" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.206783 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="99a6cea1-df62-410b-ab13-7c64f2215633" containerName="swift-ring-rebalance" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.207184 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="99a6cea1-df62-410b-ab13-7c64f2215633" containerName="swift-ring-rebalance" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.208121 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.212395 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.215261 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.236003 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l7nff"] Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.261363 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/091facbe-7485-4d1e-a813-6eb06931d862-etc-swift\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.261453 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-swiftconf\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.261512 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scwwj\" (UniqueName: \"kubernetes.io/projected/091facbe-7485-4d1e-a813-6eb06931d862-kube-api-access-scwwj\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.261553 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-ring-data-devices\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.261702 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-dispersionconf\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.261756 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-scripts\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.363685 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-ring-data-devices\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.364144 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-dispersionconf\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.364186 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-scripts\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.364297 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/091facbe-7485-4d1e-a813-6eb06931d862-etc-swift\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.364323 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-swiftconf\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.364382 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scwwj\" (UniqueName: \"kubernetes.io/projected/091facbe-7485-4d1e-a813-6eb06931d862-kube-api-access-scwwj\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.365348 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-scripts\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.365490 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/091facbe-7485-4d1e-a813-6eb06931d862-etc-swift\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.366597 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-ring-data-devices\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.371207 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-dispersionconf\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.373473 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-swiftconf\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.382400 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scwwj\" (UniqueName: \"kubernetes.io/projected/091facbe-7485-4d1e-a813-6eb06931d862-kube-api-access-scwwj\") pod \"swift-ring-rebalance-debug-l7nff\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.540943 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:11 crc kubenswrapper[4692]: I0309 09:56:11.957769 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l7nff"] Mar 09 09:56:11 crc kubenswrapper[4692]: W0309 09:56:11.963394 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod091facbe_7485_4d1e_a813_6eb06931d862.slice/crio-29f359af61c3b150e4e3954efb4039f9ca9644eda5d4bb298ac82c1ef46dcead WatchSource:0}: Error finding container 29f359af61c3b150e4e3954efb4039f9ca9644eda5d4bb298ac82c1ef46dcead: Status 404 returned error can't find the container with id 29f359af61c3b150e4e3954efb4039f9ca9644eda5d4bb298ac82c1ef46dcead Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.414454 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x6fqg"] Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.416428 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.426981 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6fqg"] Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.480051 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-utilities\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.480115 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-catalog-content\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.480138 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8j7k\" (UniqueName: \"kubernetes.io/projected/4addf7a8-446c-4074-8da5-6df32c9796f9-kube-api-access-g8j7k\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.581331 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-utilities\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.581373 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-catalog-content\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.581399 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8j7k\" (UniqueName: \"kubernetes.io/projected/4addf7a8-446c-4074-8da5-6df32c9796f9-kube-api-access-g8j7k\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.582046 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-utilities\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.582207 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-catalog-content\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.603820 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8j7k\" (UniqueName: \"kubernetes.io/projected/4addf7a8-446c-4074-8da5-6df32c9796f9-kube-api-access-g8j7k\") pod \"redhat-marketplace-x6fqg\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.704654 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" event={"ID":"091facbe-7485-4d1e-a813-6eb06931d862","Type":"ContainerStarted","Data":"3f9de65df0f5efe5792cbd77ec85413b5a264133b28c837967efbdfbbfc787ba"} Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.705183 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" event={"ID":"091facbe-7485-4d1e-a813-6eb06931d862","Type":"ContainerStarted","Data":"29f359af61c3b150e4e3954efb4039f9ca9644eda5d4bb298ac82c1ef46dcead"} Mar 09 09:56:12 crc kubenswrapper[4692]: I0309 09:56:12.741390 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:13 crc kubenswrapper[4692]: I0309 09:56:13.215401 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" podStartSLOduration=2.215368489 podStartE2EDuration="2.215368489s" podCreationTimestamp="2026-03-09 09:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:12.720273144 +0000 UTC m=+2173.545008725" watchObservedRunningTime="2026-03-09 09:56:13.215368489 +0000 UTC m=+2174.040104080" Mar 09 09:56:13 crc kubenswrapper[4692]: I0309 09:56:13.222444 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6fqg"] Mar 09 09:56:13 crc kubenswrapper[4692]: I0309 09:56:13.717972 4692 generic.go:334] "Generic (PLEG): container finished" podID="091facbe-7485-4d1e-a813-6eb06931d862" containerID="3f9de65df0f5efe5792cbd77ec85413b5a264133b28c837967efbdfbbfc787ba" exitCode=0 Mar 09 09:56:13 crc kubenswrapper[4692]: I0309 09:56:13.718091 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" event={"ID":"091facbe-7485-4d1e-a813-6eb06931d862","Type":"ContainerDied","Data":"3f9de65df0f5efe5792cbd77ec85413b5a264133b28c837967efbdfbbfc787ba"} Mar 09 09:56:13 crc kubenswrapper[4692]: I0309 09:56:13.720563 4692 generic.go:334] "Generic (PLEG): container finished" podID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerID="31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85" exitCode=0 Mar 09 09:56:13 crc kubenswrapper[4692]: I0309 09:56:13.720635 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6fqg" event={"ID":"4addf7a8-446c-4074-8da5-6df32c9796f9","Type":"ContainerDied","Data":"31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85"} Mar 09 09:56:13 crc kubenswrapper[4692]: I0309 09:56:13.721040 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6fqg" event={"ID":"4addf7a8-446c-4074-8da5-6df32c9796f9","Type":"ContainerStarted","Data":"39eae73ec7475881959e89eb2949553a03a3c0fff0d80c1152de2bde71daf0e0"} Mar 09 09:56:14 crc kubenswrapper[4692]: I0309 09:56:14.730685 4692 generic.go:334] "Generic (PLEG): container finished" podID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerID="cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0" exitCode=0 Mar 09 09:56:14 crc kubenswrapper[4692]: I0309 09:56:14.730790 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6fqg" event={"ID":"4addf7a8-446c-4074-8da5-6df32c9796f9","Type":"ContainerDied","Data":"cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0"} Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.069984 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.107882 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l7nff"] Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.116864 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l7nff"] Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.127483 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-scripts\") pod \"091facbe-7485-4d1e-a813-6eb06931d862\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.127607 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-swiftconf\") pod \"091facbe-7485-4d1e-a813-6eb06931d862\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.127651 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/091facbe-7485-4d1e-a813-6eb06931d862-etc-swift\") pod \"091facbe-7485-4d1e-a813-6eb06931d862\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.127685 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scwwj\" (UniqueName: \"kubernetes.io/projected/091facbe-7485-4d1e-a813-6eb06931d862-kube-api-access-scwwj\") pod \"091facbe-7485-4d1e-a813-6eb06931d862\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.127722 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-dispersionconf\") pod \"091facbe-7485-4d1e-a813-6eb06931d862\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.127792 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-ring-data-devices\") pod \"091facbe-7485-4d1e-a813-6eb06931d862\" (UID: \"091facbe-7485-4d1e-a813-6eb06931d862\") " Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.128752 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "091facbe-7485-4d1e-a813-6eb06931d862" (UID: "091facbe-7485-4d1e-a813-6eb06931d862"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.128820 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/091facbe-7485-4d1e-a813-6eb06931d862-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "091facbe-7485-4d1e-a813-6eb06931d862" (UID: "091facbe-7485-4d1e-a813-6eb06931d862"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.133900 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091facbe-7485-4d1e-a813-6eb06931d862-kube-api-access-scwwj" (OuterVolumeSpecName: "kube-api-access-scwwj") pod "091facbe-7485-4d1e-a813-6eb06931d862" (UID: "091facbe-7485-4d1e-a813-6eb06931d862"). InnerVolumeSpecName "kube-api-access-scwwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.150889 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-scripts" (OuterVolumeSpecName: "scripts") pod "091facbe-7485-4d1e-a813-6eb06931d862" (UID: "091facbe-7485-4d1e-a813-6eb06931d862"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.153607 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "091facbe-7485-4d1e-a813-6eb06931d862" (UID: "091facbe-7485-4d1e-a813-6eb06931d862"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.154294 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "091facbe-7485-4d1e-a813-6eb06931d862" (UID: "091facbe-7485-4d1e-a813-6eb06931d862"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.229327 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.229369 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/091facbe-7485-4d1e-a813-6eb06931d862-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.229383 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scwwj\" (UniqueName: \"kubernetes.io/projected/091facbe-7485-4d1e-a813-6eb06931d862-kube-api-access-scwwj\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.229399 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/091facbe-7485-4d1e-a813-6eb06931d862-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.229409 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.229420 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/091facbe-7485-4d1e-a813-6eb06931d862-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.740897 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l7nff" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.740896 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29f359af61c3b150e4e3954efb4039f9ca9644eda5d4bb298ac82c1ef46dcead" Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.743782 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6fqg" event={"ID":"4addf7a8-446c-4074-8da5-6df32c9796f9","Type":"ContainerStarted","Data":"43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a"} Mar 09 09:56:15 crc kubenswrapper[4692]: I0309 09:56:15.765339 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x6fqg" podStartSLOduration=2.067052837 podStartE2EDuration="3.765314838s" podCreationTimestamp="2026-03-09 09:56:12 +0000 UTC" firstStartedPulling="2026-03-09 09:56:13.722800953 +0000 UTC m=+2174.547536534" lastFinishedPulling="2026-03-09 09:56:15.421062944 +0000 UTC m=+2176.245798535" observedRunningTime="2026-03-09 09:56:15.763445825 +0000 UTC m=+2176.588181416" watchObservedRunningTime="2026-03-09 09:56:15.765314838 +0000 UTC m=+2176.590050419" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.082231 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091facbe-7485-4d1e-a813-6eb06931d862" path="/var/lib/kubelet/pods/091facbe-7485-4d1e-a813-6eb06931d862/volumes" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.257642 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wskkc"] Mar 09 09:56:16 crc kubenswrapper[4692]: E0309 09:56:16.257979 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091facbe-7485-4d1e-a813-6eb06931d862" containerName="swift-ring-rebalance" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.257992 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="091facbe-7485-4d1e-a813-6eb06931d862" containerName="swift-ring-rebalance" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.258134 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="091facbe-7485-4d1e-a813-6eb06931d862" containerName="swift-ring-rebalance" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.258598 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.261325 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.269002 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.278356 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wskkc"] Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.348576 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-ring-data-devices\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.348701 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnnrb\" (UniqueName: \"kubernetes.io/projected/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-kube-api-access-vnnrb\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.348740 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-etc-swift\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.348762 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-scripts\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.348781 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-swiftconf\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.348928 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-dispersionconf\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.451099 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnnrb\" (UniqueName: \"kubernetes.io/projected/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-kube-api-access-vnnrb\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.451649 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-etc-swift\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.451706 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-scripts\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.451731 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-swiftconf\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.451849 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-dispersionconf\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.451893 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-ring-data-devices\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.452826 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-ring-data-devices\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.453593 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-scripts\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.453819 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-etc-swift\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.457308 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-swiftconf\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.480066 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnnrb\" (UniqueName: \"kubernetes.io/projected/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-kube-api-access-vnnrb\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.480707 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-dispersionconf\") pod \"swift-ring-rebalance-debug-wskkc\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:16 crc kubenswrapper[4692]: I0309 09:56:16.581535 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:17 crc kubenswrapper[4692]: I0309 09:56:17.050191 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wskkc"] Mar 09 09:56:17 crc kubenswrapper[4692]: I0309 09:56:17.624503 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:56:17 crc kubenswrapper[4692]: I0309 09:56:17.624589 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:56:17 crc kubenswrapper[4692]: I0309 09:56:17.767813 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" event={"ID":"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7","Type":"ContainerStarted","Data":"e6a08f130ab3627ecaddd8d247938fc011b5419712a759cb11671b802d83efec"} Mar 09 09:56:17 crc kubenswrapper[4692]: I0309 09:56:17.767871 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" event={"ID":"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7","Type":"ContainerStarted","Data":"dcd23b505a7768e28b30aa3e6b5423a8d5aa4504c55d83163b7e4cb071e17192"} Mar 09 09:56:17 crc kubenswrapper[4692]: I0309 09:56:17.801880 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" podStartSLOduration=1.801854284 podStartE2EDuration="1.801854284s" podCreationTimestamp="2026-03-09 09:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:17.800923158 +0000 UTC m=+2178.625658739" watchObservedRunningTime="2026-03-09 09:56:17.801854284 +0000 UTC m=+2178.626589865" Mar 09 09:56:19 crc kubenswrapper[4692]: I0309 09:56:19.793204 4692 generic.go:334] "Generic (PLEG): container finished" podID="5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" containerID="e6a08f130ab3627ecaddd8d247938fc011b5419712a759cb11671b802d83efec" exitCode=0 Mar 09 09:56:19 crc kubenswrapper[4692]: I0309 09:56:19.793310 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" event={"ID":"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7","Type":"ContainerDied","Data":"e6a08f130ab3627ecaddd8d247938fc011b5419712a759cb11671b802d83efec"} Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.113091 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.155809 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wskkc"] Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.163230 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wskkc"] Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.240612 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-etc-swift\") pod \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.240881 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-swiftconf\") pod \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.240927 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-scripts\") pod \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.240974 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-ring-data-devices\") pod \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.241112 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnnrb\" (UniqueName: \"kubernetes.io/projected/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-kube-api-access-vnnrb\") pod \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.241145 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-dispersionconf\") pod \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\" (UID: \"5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7\") " Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.242076 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" (UID: "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.242957 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" (UID: "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.254022 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-kube-api-access-vnnrb" (OuterVolumeSpecName: "kube-api-access-vnnrb") pod "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" (UID: "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7"). InnerVolumeSpecName "kube-api-access-vnnrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.266773 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-scripts" (OuterVolumeSpecName: "scripts") pod "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" (UID: "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.271725 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" (UID: "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.276642 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" (UID: "5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.343446 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.343495 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.343514 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnnrb\" (UniqueName: \"kubernetes.io/projected/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-kube-api-access-vnnrb\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.343529 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.343540 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.343557 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.813854 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcd23b505a7768e28b30aa3e6b5423a8d5aa4504c55d83163b7e4cb071e17192" Mar 09 09:56:21 crc kubenswrapper[4692]: I0309 09:56:21.813920 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wskkc" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.100023 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" path="/var/lib/kubelet/pods/5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7/volumes" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.314132 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rttn5"] Mar 09 09:56:22 crc kubenswrapper[4692]: E0309 09:56:22.314644 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" containerName="swift-ring-rebalance" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.314667 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" containerName="swift-ring-rebalance" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.314839 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aae7e6c-56bf-4887-9f1d-3252dd1dfbc7" containerName="swift-ring-rebalance" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.315590 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.318718 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.319805 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.329286 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rttn5"] Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.359369 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-dispersionconf\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.359451 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-ring-data-devices\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.359489 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-etc-swift\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.359527 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-scripts\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.359549 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-swiftconf\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.359575 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6mrp\" (UniqueName: \"kubernetes.io/projected/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-kube-api-access-m6mrp\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.462227 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-ring-data-devices\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.462302 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-etc-swift\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.462383 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-scripts\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.462410 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-swiftconf\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.462455 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6mrp\" (UniqueName: \"kubernetes.io/projected/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-kube-api-access-m6mrp\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.462519 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-dispersionconf\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.463219 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-etc-swift\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.463719 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-scripts\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.464029 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-ring-data-devices\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.468607 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-dispersionconf\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.469671 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-swiftconf\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.486861 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6mrp\" (UniqueName: \"kubernetes.io/projected/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-kube-api-access-m6mrp\") pod \"swift-ring-rebalance-debug-rttn5\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.656002 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.742488 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.742931 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.797808 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.888341 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:22 crc kubenswrapper[4692]: I0309 09:56:22.909081 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rttn5"] Mar 09 09:56:23 crc kubenswrapper[4692]: I0309 09:56:23.044999 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6fqg"] Mar 09 09:56:23 crc kubenswrapper[4692]: I0309 09:56:23.844346 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" event={"ID":"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc","Type":"ContainerStarted","Data":"bf9b14a26aa62bffccf4424e87193645f9a38e71e5f8b1f088d70b9bc4b030d6"} Mar 09 09:56:23 crc kubenswrapper[4692]: I0309 09:56:23.844853 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" event={"ID":"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc","Type":"ContainerStarted","Data":"40fd1624141e6c93541894ab50f92f93447c4f3a50b799f91fd702715fb32849"} Mar 09 09:56:23 crc kubenswrapper[4692]: I0309 09:56:23.869905 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" podStartSLOduration=1.869880795 podStartE2EDuration="1.869880795s" podCreationTimestamp="2026-03-09 09:56:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:23.860974263 +0000 UTC m=+2184.685709844" watchObservedRunningTime="2026-03-09 09:56:23.869880795 +0000 UTC m=+2184.694616376" Mar 09 09:56:24 crc kubenswrapper[4692]: I0309 09:56:24.851643 4692 generic.go:334] "Generic (PLEG): container finished" podID="16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" containerID="bf9b14a26aa62bffccf4424e87193645f9a38e71e5f8b1f088d70b9bc4b030d6" exitCode=0 Mar 09 09:56:24 crc kubenswrapper[4692]: I0309 09:56:24.851723 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" event={"ID":"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc","Type":"ContainerDied","Data":"bf9b14a26aa62bffccf4424e87193645f9a38e71e5f8b1f088d70b9bc4b030d6"} Mar 09 09:56:24 crc kubenswrapper[4692]: I0309 09:56:24.852025 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x6fqg" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="registry-server" containerID="cri-o://43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a" gracePeriod=2 Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.268486 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.319582 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-catalog-content\") pod \"4addf7a8-446c-4074-8da5-6df32c9796f9\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.319645 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-utilities\") pod \"4addf7a8-446c-4074-8da5-6df32c9796f9\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.319781 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8j7k\" (UniqueName: \"kubernetes.io/projected/4addf7a8-446c-4074-8da5-6df32c9796f9-kube-api-access-g8j7k\") pod \"4addf7a8-446c-4074-8da5-6df32c9796f9\" (UID: \"4addf7a8-446c-4074-8da5-6df32c9796f9\") " Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.320496 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-utilities" (OuterVolumeSpecName: "utilities") pod "4addf7a8-446c-4074-8da5-6df32c9796f9" (UID: "4addf7a8-446c-4074-8da5-6df32c9796f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.326739 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4addf7a8-446c-4074-8da5-6df32c9796f9-kube-api-access-g8j7k" (OuterVolumeSpecName: "kube-api-access-g8j7k") pod "4addf7a8-446c-4074-8da5-6df32c9796f9" (UID: "4addf7a8-446c-4074-8da5-6df32c9796f9"). InnerVolumeSpecName "kube-api-access-g8j7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.349906 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4addf7a8-446c-4074-8da5-6df32c9796f9" (UID: "4addf7a8-446c-4074-8da5-6df32c9796f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.422637 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.422688 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4addf7a8-446c-4074-8da5-6df32c9796f9-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.422700 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8j7k\" (UniqueName: \"kubernetes.io/projected/4addf7a8-446c-4074-8da5-6df32c9796f9-kube-api-access-g8j7k\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.862242 4692 generic.go:334] "Generic (PLEG): container finished" podID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerID="43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a" exitCode=0 Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.862338 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6fqg" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.862337 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6fqg" event={"ID":"4addf7a8-446c-4074-8da5-6df32c9796f9","Type":"ContainerDied","Data":"43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a"} Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.863469 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6fqg" event={"ID":"4addf7a8-446c-4074-8da5-6df32c9796f9","Type":"ContainerDied","Data":"39eae73ec7475881959e89eb2949553a03a3c0fff0d80c1152de2bde71daf0e0"} Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.863514 4692 scope.go:117] "RemoveContainer" containerID="43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.886190 4692 scope.go:117] "RemoveContainer" containerID="cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.932602 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6fqg"] Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.929167 4692 scope.go:117] "RemoveContainer" containerID="31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.941075 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6fqg"] Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.959645 4692 scope.go:117] "RemoveContainer" containerID="43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a" Mar 09 09:56:25 crc kubenswrapper[4692]: E0309 09:56:25.960397 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a\": container with ID starting with 43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a not found: ID does not exist" containerID="43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.960462 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a"} err="failed to get container status \"43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a\": rpc error: code = NotFound desc = could not find container \"43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a\": container with ID starting with 43531119ea1abed29bcc0a75936fb9d6e018076d9eca373a3ba0f89a31e24f9a not found: ID does not exist" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.960504 4692 scope.go:117] "RemoveContainer" containerID="cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0" Mar 09 09:56:25 crc kubenswrapper[4692]: E0309 09:56:25.960823 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0\": container with ID starting with cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0 not found: ID does not exist" containerID="cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.960849 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0"} err="failed to get container status \"cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0\": rpc error: code = NotFound desc = could not find container \"cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0\": container with ID starting with cd85619b0c9c25c2f53c2e013cccec667086f2600587aee64738e56c39d65fa0 not found: ID does not exist" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.960865 4692 scope.go:117] "RemoveContainer" containerID="31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85" Mar 09 09:56:25 crc kubenswrapper[4692]: E0309 09:56:25.961220 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85\": container with ID starting with 31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85 not found: ID does not exist" containerID="31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85" Mar 09 09:56:25 crc kubenswrapper[4692]: I0309 09:56:25.961243 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85"} err="failed to get container status \"31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85\": rpc error: code = NotFound desc = could not find container \"31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85\": container with ID starting with 31eef484af17ca6cd8959a2dcb28518ee2212b0a772ba8936de51f28785a4c85 not found: ID does not exist" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.084786 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" path="/var/lib/kubelet/pods/4addf7a8-446c-4074-8da5-6df32c9796f9/volumes" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.159187 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.201828 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rttn5"] Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.208799 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rttn5"] Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.241823 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-scripts\") pod \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.241943 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-ring-data-devices\") pod \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.242011 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6mrp\" (UniqueName: \"kubernetes.io/projected/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-kube-api-access-m6mrp\") pod \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.242144 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-dispersionconf\") pod \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.242235 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-swiftconf\") pod \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.242307 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-etc-swift\") pod \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\" (UID: \"16bf4f53-e1a0-43b4-8e06-c3be9e844dbc\") " Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.242768 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" (UID: "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.243520 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" (UID: "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.247116 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-kube-api-access-m6mrp" (OuterVolumeSpecName: "kube-api-access-m6mrp") pod "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" (UID: "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc"). InnerVolumeSpecName "kube-api-access-m6mrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.268549 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" (UID: "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.269735 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" (UID: "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.276802 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-scripts" (OuterVolumeSpecName: "scripts") pod "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" (UID: "16bf4f53-e1a0-43b4-8e06-c3be9e844dbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.344707 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.344757 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.344787 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.344800 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.344813 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6mrp\" (UniqueName: \"kubernetes.io/projected/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-kube-api-access-m6mrp\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.344828 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.876909 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40fd1624141e6c93541894ab50f92f93447c4f3a50b799f91fd702715fb32849" Mar 09 09:56:26 crc kubenswrapper[4692]: I0309 09:56:26.877100 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rttn5" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.339086 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk"] Mar 09 09:56:27 crc kubenswrapper[4692]: E0309 09:56:27.339455 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="extract-content" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.339478 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="extract-content" Mar 09 09:56:27 crc kubenswrapper[4692]: E0309 09:56:27.339495 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="extract-utilities" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.339504 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="extract-utilities" Mar 09 09:56:27 crc kubenswrapper[4692]: E0309 09:56:27.339514 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" containerName="swift-ring-rebalance" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.339523 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" containerName="swift-ring-rebalance" Mar 09 09:56:27 crc kubenswrapper[4692]: E0309 09:56:27.339539 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="registry-server" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.339547 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="registry-server" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.339708 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="4addf7a8-446c-4074-8da5-6df32c9796f9" containerName="registry-server" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.339730 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" containerName="swift-ring-rebalance" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.340262 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.342975 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.344567 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.361565 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk"] Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.362319 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-swiftconf\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.366870 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmc4c\" (UniqueName: \"kubernetes.io/projected/99c82069-0db4-4d26-81b3-22708c05b7c4-kube-api-access-wmc4c\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.367008 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-scripts\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.367120 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-dispersionconf\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.367202 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-ring-data-devices\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.367496 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99c82069-0db4-4d26-81b3-22708c05b7c4-etc-swift\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.469108 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-scripts\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.469237 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-dispersionconf\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.469265 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-ring-data-devices\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.469341 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99c82069-0db4-4d26-81b3-22708c05b7c4-etc-swift\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.469394 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-swiftconf\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.469468 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmc4c\" (UniqueName: \"kubernetes.io/projected/99c82069-0db4-4d26-81b3-22708c05b7c4-kube-api-access-wmc4c\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.470464 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99c82069-0db4-4d26-81b3-22708c05b7c4-etc-swift\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.471644 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-ring-data-devices\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.471697 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-scripts\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.477182 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-swiftconf\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.485802 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-dispersionconf\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.493144 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmc4c\" (UniqueName: \"kubernetes.io/projected/99c82069-0db4-4d26-81b3-22708c05b7c4-kube-api-access-wmc4c\") pod \"swift-ring-rebalance-debug-d5bnk\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:27 crc kubenswrapper[4692]: I0309 09:56:27.667148 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:28 crc kubenswrapper[4692]: I0309 09:56:28.085395 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16bf4f53-e1a0-43b4-8e06-c3be9e844dbc" path="/var/lib/kubelet/pods/16bf4f53-e1a0-43b4-8e06-c3be9e844dbc/volumes" Mar 09 09:56:28 crc kubenswrapper[4692]: I0309 09:56:28.109489 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk"] Mar 09 09:56:28 crc kubenswrapper[4692]: I0309 09:56:28.914666 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" event={"ID":"99c82069-0db4-4d26-81b3-22708c05b7c4","Type":"ContainerStarted","Data":"14a05f337e0b5bd94fb5dd17b0c0e9bfc8a1552e5e92cd8b64220ba13a73e091"} Mar 09 09:56:28 crc kubenswrapper[4692]: I0309 09:56:28.915009 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" event={"ID":"99c82069-0db4-4d26-81b3-22708c05b7c4","Type":"ContainerStarted","Data":"26e4d60967977c667cd71f07d9ac8b86586cf4068b00d5ca4348e610a7e7d22c"} Mar 09 09:56:28 crc kubenswrapper[4692]: I0309 09:56:28.933609 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" podStartSLOduration=1.933577638 podStartE2EDuration="1.933577638s" podCreationTimestamp="2026-03-09 09:56:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:28.933533967 +0000 UTC m=+2189.758269558" watchObservedRunningTime="2026-03-09 09:56:28.933577638 +0000 UTC m=+2189.758313229" Mar 09 09:56:29 crc kubenswrapper[4692]: I0309 09:56:29.923832 4692 generic.go:334] "Generic (PLEG): container finished" podID="99c82069-0db4-4d26-81b3-22708c05b7c4" containerID="14a05f337e0b5bd94fb5dd17b0c0e9bfc8a1552e5e92cd8b64220ba13a73e091" exitCode=0 Mar 09 09:56:29 crc kubenswrapper[4692]: I0309 09:56:29.923951 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" event={"ID":"99c82069-0db4-4d26-81b3-22708c05b7c4","Type":"ContainerDied","Data":"14a05f337e0b5bd94fb5dd17b0c0e9bfc8a1552e5e92cd8b64220ba13a73e091"} Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.185308 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.221092 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk"] Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.229443 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk"] Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.240721 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmc4c\" (UniqueName: \"kubernetes.io/projected/99c82069-0db4-4d26-81b3-22708c05b7c4-kube-api-access-wmc4c\") pod \"99c82069-0db4-4d26-81b3-22708c05b7c4\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.240839 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-swiftconf\") pod \"99c82069-0db4-4d26-81b3-22708c05b7c4\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.240893 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-scripts\") pod \"99c82069-0db4-4d26-81b3-22708c05b7c4\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.240966 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-dispersionconf\") pod \"99c82069-0db4-4d26-81b3-22708c05b7c4\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.241015 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99c82069-0db4-4d26-81b3-22708c05b7c4-etc-swift\") pod \"99c82069-0db4-4d26-81b3-22708c05b7c4\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.241093 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-ring-data-devices\") pod \"99c82069-0db4-4d26-81b3-22708c05b7c4\" (UID: \"99c82069-0db4-4d26-81b3-22708c05b7c4\") " Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.242090 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "99c82069-0db4-4d26-81b3-22708c05b7c4" (UID: "99c82069-0db4-4d26-81b3-22708c05b7c4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.242327 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99c82069-0db4-4d26-81b3-22708c05b7c4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "99c82069-0db4-4d26-81b3-22708c05b7c4" (UID: "99c82069-0db4-4d26-81b3-22708c05b7c4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.246338 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c82069-0db4-4d26-81b3-22708c05b7c4-kube-api-access-wmc4c" (OuterVolumeSpecName: "kube-api-access-wmc4c") pod "99c82069-0db4-4d26-81b3-22708c05b7c4" (UID: "99c82069-0db4-4d26-81b3-22708c05b7c4"). InnerVolumeSpecName "kube-api-access-wmc4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.260561 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-scripts" (OuterVolumeSpecName: "scripts") pod "99c82069-0db4-4d26-81b3-22708c05b7c4" (UID: "99c82069-0db4-4d26-81b3-22708c05b7c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.263298 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "99c82069-0db4-4d26-81b3-22708c05b7c4" (UID: "99c82069-0db4-4d26-81b3-22708c05b7c4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.264419 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "99c82069-0db4-4d26-81b3-22708c05b7c4" (UID: "99c82069-0db4-4d26-81b3-22708c05b7c4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.342974 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.343292 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/99c82069-0db4-4d26-81b3-22708c05b7c4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.343357 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.343415 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmc4c\" (UniqueName: \"kubernetes.io/projected/99c82069-0db4-4d26-81b3-22708c05b7c4-kube-api-access-wmc4c\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.343480 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/99c82069-0db4-4d26-81b3-22708c05b7c4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.343532 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c82069-0db4-4d26-81b3-22708c05b7c4-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.940969 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26e4d60967977c667cd71f07d9ac8b86586cf4068b00d5ca4348e610a7e7d22c" Mar 09 09:56:31 crc kubenswrapper[4692]: I0309 09:56:31.941286 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-d5bnk" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.080671 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c82069-0db4-4d26-81b3-22708c05b7c4" path="/var/lib/kubelet/pods/99c82069-0db4-4d26-81b3-22708c05b7c4/volumes" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.361674 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq"] Mar 09 09:56:32 crc kubenswrapper[4692]: E0309 09:56:32.361962 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c82069-0db4-4d26-81b3-22708c05b7c4" containerName="swift-ring-rebalance" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.361974 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c82069-0db4-4d26-81b3-22708c05b7c4" containerName="swift-ring-rebalance" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.362124 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c82069-0db4-4d26-81b3-22708c05b7c4" containerName="swift-ring-rebalance" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.362686 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.365088 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.366517 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.374896 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq"] Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.458295 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-ring-data-devices\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.458389 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9qvp\" (UniqueName: \"kubernetes.io/projected/2f716225-6604-41c2-90d0-0a5061156010-kube-api-access-x9qvp\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.458472 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-dispersionconf\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.458495 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-swiftconf\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.458522 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-scripts\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.458550 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f716225-6604-41c2-90d0-0a5061156010-etc-swift\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.560039 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9qvp\" (UniqueName: \"kubernetes.io/projected/2f716225-6604-41c2-90d0-0a5061156010-kube-api-access-x9qvp\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.560101 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-dispersionconf\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.560118 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-swiftconf\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.560152 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-scripts\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.560202 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f716225-6604-41c2-90d0-0a5061156010-etc-swift\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.560235 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-ring-data-devices\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.560960 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-ring-data-devices\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.561038 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f716225-6604-41c2-90d0-0a5061156010-etc-swift\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.561274 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-scripts\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.567425 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-swiftconf\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.568531 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-dispersionconf\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.576871 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9qvp\" (UniqueName: \"kubernetes.io/projected/2f716225-6604-41c2-90d0-0a5061156010-kube-api-access-x9qvp\") pod \"swift-ring-rebalance-debug-cl9bq\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:32 crc kubenswrapper[4692]: I0309 09:56:32.688435 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:33 crc kubenswrapper[4692]: I0309 09:56:33.086686 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq"] Mar 09 09:56:33 crc kubenswrapper[4692]: W0309 09:56:33.091843 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f716225_6604_41c2_90d0_0a5061156010.slice/crio-dd7a949f4599dba833d230cd513225b644ba8d4cf83a620bd0088d811ed569d7 WatchSource:0}: Error finding container dd7a949f4599dba833d230cd513225b644ba8d4cf83a620bd0088d811ed569d7: Status 404 returned error can't find the container with id dd7a949f4599dba833d230cd513225b644ba8d4cf83a620bd0088d811ed569d7 Mar 09 09:56:33 crc kubenswrapper[4692]: I0309 09:56:33.958058 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" event={"ID":"2f716225-6604-41c2-90d0-0a5061156010","Type":"ContainerStarted","Data":"75558d2113ea0b9d568a4ff68e63391d57278a629ac5b42eaa03b5eff4c8d930"} Mar 09 09:56:33 crc kubenswrapper[4692]: I0309 09:56:33.958422 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" event={"ID":"2f716225-6604-41c2-90d0-0a5061156010","Type":"ContainerStarted","Data":"dd7a949f4599dba833d230cd513225b644ba8d4cf83a620bd0088d811ed569d7"} Mar 09 09:56:33 crc kubenswrapper[4692]: I0309 09:56:33.976493 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" podStartSLOduration=1.976468091 podStartE2EDuration="1.976468091s" podCreationTimestamp="2026-03-09 09:56:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:33.973145307 +0000 UTC m=+2194.797880918" watchObservedRunningTime="2026-03-09 09:56:33.976468091 +0000 UTC m=+2194.801203682" Mar 09 09:56:34 crc kubenswrapper[4692]: I0309 09:56:34.985312 4692 generic.go:334] "Generic (PLEG): container finished" podID="2f716225-6604-41c2-90d0-0a5061156010" containerID="75558d2113ea0b9d568a4ff68e63391d57278a629ac5b42eaa03b5eff4c8d930" exitCode=0 Mar 09 09:56:34 crc kubenswrapper[4692]: I0309 09:56:34.985496 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" event={"ID":"2f716225-6604-41c2-90d0-0a5061156010","Type":"ContainerDied","Data":"75558d2113ea0b9d568a4ff68e63391d57278a629ac5b42eaa03b5eff4c8d930"} Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.292671 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.314687 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f716225-6604-41c2-90d0-0a5061156010-etc-swift\") pod \"2f716225-6604-41c2-90d0-0a5061156010\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.314736 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-dispersionconf\") pod \"2f716225-6604-41c2-90d0-0a5061156010\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.314765 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-swiftconf\") pod \"2f716225-6604-41c2-90d0-0a5061156010\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.314795 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-scripts\") pod \"2f716225-6604-41c2-90d0-0a5061156010\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.314833 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9qvp\" (UniqueName: \"kubernetes.io/projected/2f716225-6604-41c2-90d0-0a5061156010-kube-api-access-x9qvp\") pod \"2f716225-6604-41c2-90d0-0a5061156010\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.314870 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-ring-data-devices\") pod \"2f716225-6604-41c2-90d0-0a5061156010\" (UID: \"2f716225-6604-41c2-90d0-0a5061156010\") " Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.315651 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2f716225-6604-41c2-90d0-0a5061156010" (UID: "2f716225-6604-41c2-90d0-0a5061156010"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.316503 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f716225-6604-41c2-90d0-0a5061156010-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2f716225-6604-41c2-90d0-0a5061156010" (UID: "2f716225-6604-41c2-90d0-0a5061156010"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.320679 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f716225-6604-41c2-90d0-0a5061156010-kube-api-access-x9qvp" (OuterVolumeSpecName: "kube-api-access-x9qvp") pod "2f716225-6604-41c2-90d0-0a5061156010" (UID: "2f716225-6604-41c2-90d0-0a5061156010"). InnerVolumeSpecName "kube-api-access-x9qvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.338569 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2f716225-6604-41c2-90d0-0a5061156010" (UID: "2f716225-6604-41c2-90d0-0a5061156010"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.338750 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-scripts" (OuterVolumeSpecName: "scripts") pod "2f716225-6604-41c2-90d0-0a5061156010" (UID: "2f716225-6604-41c2-90d0-0a5061156010"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.341524 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2f716225-6604-41c2-90d0-0a5061156010" (UID: "2f716225-6604-41c2-90d0-0a5061156010"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.350944 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq"] Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.361888 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq"] Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.416197 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.416265 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.416285 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9qvp\" (UniqueName: \"kubernetes.io/projected/2f716225-6604-41c2-90d0-0a5061156010-kube-api-access-x9qvp\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.416297 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2f716225-6604-41c2-90d0-0a5061156010-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.416307 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2f716225-6604-41c2-90d0-0a5061156010-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:36 crc kubenswrapper[4692]: I0309 09:56:36.416315 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2f716225-6604-41c2-90d0-0a5061156010-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.019122 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd7a949f4599dba833d230cd513225b644ba8d4cf83a620bd0088d811ed569d7" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.019175 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-cl9bq" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.468323 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l6t69"] Mar 09 09:56:37 crc kubenswrapper[4692]: E0309 09:56:37.469287 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f716225-6604-41c2-90d0-0a5061156010" containerName="swift-ring-rebalance" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.469304 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f716225-6604-41c2-90d0-0a5061156010" containerName="swift-ring-rebalance" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.469459 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f716225-6604-41c2-90d0-0a5061156010" containerName="swift-ring-rebalance" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.470054 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.472776 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.474453 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.480263 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l6t69"] Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.532290 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-swiftconf\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.532327 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-dispersionconf\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.532351 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-scripts\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.532400 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-ring-data-devices\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.532629 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfdmn\" (UniqueName: \"kubernetes.io/projected/f76999f2-2c27-4253-bdf5-9837706acb05-kube-api-access-nfdmn\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.532743 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f76999f2-2c27-4253-bdf5-9837706acb05-etc-swift\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.634100 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-swiftconf\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.634142 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-scripts\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.634207 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-dispersionconf\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.634286 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-ring-data-devices\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.634319 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfdmn\" (UniqueName: \"kubernetes.io/projected/f76999f2-2c27-4253-bdf5-9837706acb05-kube-api-access-nfdmn\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.634358 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f76999f2-2c27-4253-bdf5-9837706acb05-etc-swift\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.635073 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f76999f2-2c27-4253-bdf5-9837706acb05-etc-swift\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.635574 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-scripts\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.635669 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-ring-data-devices\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.639635 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-dispersionconf\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.642000 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-swiftconf\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.652966 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfdmn\" (UniqueName: \"kubernetes.io/projected/f76999f2-2c27-4253-bdf5-9837706acb05-kube-api-access-nfdmn\") pod \"swift-ring-rebalance-debug-l6t69\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:37 crc kubenswrapper[4692]: I0309 09:56:37.794032 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:38 crc kubenswrapper[4692]: I0309 09:56:38.082845 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f716225-6604-41c2-90d0-0a5061156010" path="/var/lib/kubelet/pods/2f716225-6604-41c2-90d0-0a5061156010/volumes" Mar 09 09:56:38 crc kubenswrapper[4692]: I0309 09:56:38.207578 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l6t69"] Mar 09 09:56:39 crc kubenswrapper[4692]: I0309 09:56:39.039251 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" event={"ID":"f76999f2-2c27-4253-bdf5-9837706acb05","Type":"ContainerStarted","Data":"05a0ed0b9239c5a98893eb817303d2b523623665c94397ec68996938dcbadd5f"} Mar 09 09:56:39 crc kubenswrapper[4692]: I0309 09:56:39.039678 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" event={"ID":"f76999f2-2c27-4253-bdf5-9837706acb05","Type":"ContainerStarted","Data":"c83c231c8b880f10a4b632eca2cbed955e2d94ba600e3c2c902c97863cf25146"} Mar 09 09:56:39 crc kubenswrapper[4692]: I0309 09:56:39.058796 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" podStartSLOduration=2.05876143 podStartE2EDuration="2.05876143s" podCreationTimestamp="2026-03-09 09:56:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:39.055450676 +0000 UTC m=+2199.880186257" watchObservedRunningTime="2026-03-09 09:56:39.05876143 +0000 UTC m=+2199.883497021" Mar 09 09:56:40 crc kubenswrapper[4692]: I0309 09:56:40.050276 4692 generic.go:334] "Generic (PLEG): container finished" podID="f76999f2-2c27-4253-bdf5-9837706acb05" containerID="05a0ed0b9239c5a98893eb817303d2b523623665c94397ec68996938dcbadd5f" exitCode=0 Mar 09 09:56:40 crc kubenswrapper[4692]: I0309 09:56:40.050839 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" event={"ID":"f76999f2-2c27-4253-bdf5-9837706acb05","Type":"ContainerDied","Data":"05a0ed0b9239c5a98893eb817303d2b523623665c94397ec68996938dcbadd5f"} Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.384529 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.420986 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l6t69"] Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.425547 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-l6t69"] Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.493344 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-ring-data-devices\") pod \"f76999f2-2c27-4253-bdf5-9837706acb05\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.493422 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfdmn\" (UniqueName: \"kubernetes.io/projected/f76999f2-2c27-4253-bdf5-9837706acb05-kube-api-access-nfdmn\") pod \"f76999f2-2c27-4253-bdf5-9837706acb05\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.493503 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-dispersionconf\") pod \"f76999f2-2c27-4253-bdf5-9837706acb05\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.493539 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f76999f2-2c27-4253-bdf5-9837706acb05-etc-swift\") pod \"f76999f2-2c27-4253-bdf5-9837706acb05\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.493580 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-swiftconf\") pod \"f76999f2-2c27-4253-bdf5-9837706acb05\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.493632 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-scripts\") pod \"f76999f2-2c27-4253-bdf5-9837706acb05\" (UID: \"f76999f2-2c27-4253-bdf5-9837706acb05\") " Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.493936 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f76999f2-2c27-4253-bdf5-9837706acb05" (UID: "f76999f2-2c27-4253-bdf5-9837706acb05"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.494565 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f76999f2-2c27-4253-bdf5-9837706acb05-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f76999f2-2c27-4253-bdf5-9837706acb05" (UID: "f76999f2-2c27-4253-bdf5-9837706acb05"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.501323 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f76999f2-2c27-4253-bdf5-9837706acb05-kube-api-access-nfdmn" (OuterVolumeSpecName: "kube-api-access-nfdmn") pod "f76999f2-2c27-4253-bdf5-9837706acb05" (UID: "f76999f2-2c27-4253-bdf5-9837706acb05"). InnerVolumeSpecName "kube-api-access-nfdmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.519934 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-scripts" (OuterVolumeSpecName: "scripts") pod "f76999f2-2c27-4253-bdf5-9837706acb05" (UID: "f76999f2-2c27-4253-bdf5-9837706acb05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.523542 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f76999f2-2c27-4253-bdf5-9837706acb05" (UID: "f76999f2-2c27-4253-bdf5-9837706acb05"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.524133 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f76999f2-2c27-4253-bdf5-9837706acb05" (UID: "f76999f2-2c27-4253-bdf5-9837706acb05"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.595992 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.596592 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f76999f2-2c27-4253-bdf5-9837706acb05-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.596607 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f76999f2-2c27-4253-bdf5-9837706acb05-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.596620 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.596634 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f76999f2-2c27-4253-bdf5-9837706acb05-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:41 crc kubenswrapper[4692]: I0309 09:56:41.596651 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfdmn\" (UniqueName: \"kubernetes.io/projected/f76999f2-2c27-4253-bdf5-9837706acb05-kube-api-access-nfdmn\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.089048 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-l6t69" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.089749 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f76999f2-2c27-4253-bdf5-9837706acb05" path="/var/lib/kubelet/pods/f76999f2-2c27-4253-bdf5-9837706acb05/volumes" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.090409 4692 scope.go:117] "RemoveContainer" containerID="05a0ed0b9239c5a98893eb817303d2b523623665c94397ec68996938dcbadd5f" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.583571 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-svtb8"] Mar 09 09:56:42 crc kubenswrapper[4692]: E0309 09:56:42.583875 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f76999f2-2c27-4253-bdf5-9837706acb05" containerName="swift-ring-rebalance" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.583890 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f76999f2-2c27-4253-bdf5-9837706acb05" containerName="swift-ring-rebalance" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.584069 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f76999f2-2c27-4253-bdf5-9837706acb05" containerName="swift-ring-rebalance" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.584608 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.587054 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.587254 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.595946 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-svtb8"] Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.717067 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-swiftconf\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.717116 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-dispersionconf\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.717150 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-ring-data-devices\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.717185 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-scripts\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.717236 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khd64\" (UniqueName: \"kubernetes.io/projected/e142c3a5-b983-47fd-a29b-1365f665474f-kube-api-access-khd64\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.717259 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e142c3a5-b983-47fd-a29b-1365f665474f-etc-swift\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818102 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khd64\" (UniqueName: \"kubernetes.io/projected/e142c3a5-b983-47fd-a29b-1365f665474f-kube-api-access-khd64\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818151 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e142c3a5-b983-47fd-a29b-1365f665474f-etc-swift\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818233 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-swiftconf\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818255 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-dispersionconf\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818282 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-ring-data-devices\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818304 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-scripts\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818600 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e142c3a5-b983-47fd-a29b-1365f665474f-etc-swift\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.818951 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-scripts\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.819250 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-ring-data-devices\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.824899 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-dispersionconf\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.825456 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-swiftconf\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.840327 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khd64\" (UniqueName: \"kubernetes.io/projected/e142c3a5-b983-47fd-a29b-1365f665474f-kube-api-access-khd64\") pod \"swift-ring-rebalance-debug-svtb8\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:42 crc kubenswrapper[4692]: I0309 09:56:42.917060 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:43 crc kubenswrapper[4692]: I0309 09:56:43.374820 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-svtb8"] Mar 09 09:56:44 crc kubenswrapper[4692]: I0309 09:56:44.119845 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" event={"ID":"e142c3a5-b983-47fd-a29b-1365f665474f","Type":"ContainerStarted","Data":"1096f6b01f087e07b51c13627e1b1c3df945b3194a1b7183c99dcbc8326a04e4"} Mar 09 09:56:44 crc kubenswrapper[4692]: I0309 09:56:44.120179 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" event={"ID":"e142c3a5-b983-47fd-a29b-1365f665474f","Type":"ContainerStarted","Data":"c1ee0c72ea1d6fae36209e1edd977934304328bc73b40df84d669e2c0ac1d46f"} Mar 09 09:56:44 crc kubenswrapper[4692]: I0309 09:56:44.141123 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" podStartSLOduration=2.141077579 podStartE2EDuration="2.141077579s" podCreationTimestamp="2026-03-09 09:56:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:44.13579476 +0000 UTC m=+2204.960530361" watchObservedRunningTime="2026-03-09 09:56:44.141077579 +0000 UTC m=+2204.965813200" Mar 09 09:56:45 crc kubenswrapper[4692]: I0309 09:56:45.137721 4692 generic.go:334] "Generic (PLEG): container finished" podID="e142c3a5-b983-47fd-a29b-1365f665474f" containerID="1096f6b01f087e07b51c13627e1b1c3df945b3194a1b7183c99dcbc8326a04e4" exitCode=0 Mar 09 09:56:45 crc kubenswrapper[4692]: I0309 09:56:45.138110 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" event={"ID":"e142c3a5-b983-47fd-a29b-1365f665474f","Type":"ContainerDied","Data":"1096f6b01f087e07b51c13627e1b1c3df945b3194a1b7183c99dcbc8326a04e4"} Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.428719 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.469909 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-svtb8"] Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.475758 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-svtb8"] Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.589410 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-ring-data-devices\") pod \"e142c3a5-b983-47fd-a29b-1365f665474f\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.589558 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khd64\" (UniqueName: \"kubernetes.io/projected/e142c3a5-b983-47fd-a29b-1365f665474f-kube-api-access-khd64\") pod \"e142c3a5-b983-47fd-a29b-1365f665474f\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.589600 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e142c3a5-b983-47fd-a29b-1365f665474f-etc-swift\") pod \"e142c3a5-b983-47fd-a29b-1365f665474f\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.589638 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-scripts\") pod \"e142c3a5-b983-47fd-a29b-1365f665474f\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.589671 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-dispersionconf\") pod \"e142c3a5-b983-47fd-a29b-1365f665474f\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.589693 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-swiftconf\") pod \"e142c3a5-b983-47fd-a29b-1365f665474f\" (UID: \"e142c3a5-b983-47fd-a29b-1365f665474f\") " Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.590060 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e142c3a5-b983-47fd-a29b-1365f665474f" (UID: "e142c3a5-b983-47fd-a29b-1365f665474f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.590341 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e142c3a5-b983-47fd-a29b-1365f665474f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e142c3a5-b983-47fd-a29b-1365f665474f" (UID: "e142c3a5-b983-47fd-a29b-1365f665474f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.591632 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.591669 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e142c3a5-b983-47fd-a29b-1365f665474f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.598637 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e142c3a5-b983-47fd-a29b-1365f665474f-kube-api-access-khd64" (OuterVolumeSpecName: "kube-api-access-khd64") pod "e142c3a5-b983-47fd-a29b-1365f665474f" (UID: "e142c3a5-b983-47fd-a29b-1365f665474f"). InnerVolumeSpecName "kube-api-access-khd64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.614645 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e142c3a5-b983-47fd-a29b-1365f665474f" (UID: "e142c3a5-b983-47fd-a29b-1365f665474f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.616574 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-scripts" (OuterVolumeSpecName: "scripts") pod "e142c3a5-b983-47fd-a29b-1365f665474f" (UID: "e142c3a5-b983-47fd-a29b-1365f665474f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.616771 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e142c3a5-b983-47fd-a29b-1365f665474f" (UID: "e142c3a5-b983-47fd-a29b-1365f665474f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.692558 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khd64\" (UniqueName: \"kubernetes.io/projected/e142c3a5-b983-47fd-a29b-1365f665474f-kube-api-access-khd64\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.692633 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e142c3a5-b983-47fd-a29b-1365f665474f-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.692650 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:46 crc kubenswrapper[4692]: I0309 09:56:46.692668 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e142c3a5-b983-47fd-a29b-1365f665474f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.164363 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1ee0c72ea1d6fae36209e1edd977934304328bc73b40df84d669e2c0ac1d46f" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.164428 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-svtb8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.603205 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8"] Mar 09 09:56:47 crc kubenswrapper[4692]: E0309 09:56:47.604749 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e142c3a5-b983-47fd-a29b-1365f665474f" containerName="swift-ring-rebalance" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.604831 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e142c3a5-b983-47fd-a29b-1365f665474f" containerName="swift-ring-rebalance" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.605071 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e142c3a5-b983-47fd-a29b-1365f665474f" containerName="swift-ring-rebalance" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.605578 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.607592 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.607591 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.618260 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8"] Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.623788 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.623934 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.704388 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-dispersionconf\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.704466 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-swiftconf\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.704492 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-ring-data-devices\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.704522 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-scripts\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.704582 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d3320db-5949-4f67-8c9b-b709e2e7e542-etc-swift\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.704625 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w282\" (UniqueName: \"kubernetes.io/projected/5d3320db-5949-4f67-8c9b-b709e2e7e542-kube-api-access-6w282\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.805601 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-swiftconf\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.805671 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-ring-data-devices\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.805710 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-scripts\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.805761 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d3320db-5949-4f67-8c9b-b709e2e7e542-etc-swift\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.805817 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w282\" (UniqueName: \"kubernetes.io/projected/5d3320db-5949-4f67-8c9b-b709e2e7e542-kube-api-access-6w282\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.805847 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-dispersionconf\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.806404 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d3320db-5949-4f67-8c9b-b709e2e7e542-etc-swift\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.806617 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-ring-data-devices\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.807108 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-scripts\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.811485 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-swiftconf\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.820679 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-dispersionconf\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.822139 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w282\" (UniqueName: \"kubernetes.io/projected/5d3320db-5949-4f67-8c9b-b709e2e7e542-kube-api-access-6w282\") pod \"swift-ring-rebalance-debug-ms8s8\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:47 crc kubenswrapper[4692]: I0309 09:56:47.924036 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:48 crc kubenswrapper[4692]: I0309 09:56:48.081566 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e142c3a5-b983-47fd-a29b-1365f665474f" path="/var/lib/kubelet/pods/e142c3a5-b983-47fd-a29b-1365f665474f/volumes" Mar 09 09:56:48 crc kubenswrapper[4692]: I0309 09:56:48.377222 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8"] Mar 09 09:56:48 crc kubenswrapper[4692]: W0309 09:56:48.382835 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d3320db_5949_4f67_8c9b_b709e2e7e542.slice/crio-0a77ca3e06012e60d7f05ac70ca884efc8fa83c9cbb5029c5ad4adf7ba1b8718 WatchSource:0}: Error finding container 0a77ca3e06012e60d7f05ac70ca884efc8fa83c9cbb5029c5ad4adf7ba1b8718: Status 404 returned error can't find the container with id 0a77ca3e06012e60d7f05ac70ca884efc8fa83c9cbb5029c5ad4adf7ba1b8718 Mar 09 09:56:49 crc kubenswrapper[4692]: I0309 09:56:49.185278 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" event={"ID":"5d3320db-5949-4f67-8c9b-b709e2e7e542","Type":"ContainerStarted","Data":"4d7b7a3ddcaea2c88415b27c6976b57b0ca39cda081ce8e68d3ff0496996b830"} Mar 09 09:56:49 crc kubenswrapper[4692]: I0309 09:56:49.185376 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" event={"ID":"5d3320db-5949-4f67-8c9b-b709e2e7e542","Type":"ContainerStarted","Data":"0a77ca3e06012e60d7f05ac70ca884efc8fa83c9cbb5029c5ad4adf7ba1b8718"} Mar 09 09:56:49 crc kubenswrapper[4692]: I0309 09:56:49.215127 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" podStartSLOduration=2.215080966 podStartE2EDuration="2.215080966s" podCreationTimestamp="2026-03-09 09:56:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:49.201054809 +0000 UTC m=+2210.025790410" watchObservedRunningTime="2026-03-09 09:56:49.215080966 +0000 UTC m=+2210.039816577" Mar 09 09:56:50 crc kubenswrapper[4692]: I0309 09:56:50.197785 4692 generic.go:334] "Generic (PLEG): container finished" podID="5d3320db-5949-4f67-8c9b-b709e2e7e542" containerID="4d7b7a3ddcaea2c88415b27c6976b57b0ca39cda081ce8e68d3ff0496996b830" exitCode=0 Mar 09 09:56:50 crc kubenswrapper[4692]: I0309 09:56:50.197893 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" event={"ID":"5d3320db-5949-4f67-8c9b-b709e2e7e542","Type":"ContainerDied","Data":"4d7b7a3ddcaea2c88415b27c6976b57b0ca39cda081ce8e68d3ff0496996b830"} Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.514358 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.555039 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8"] Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.557830 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8"] Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.596898 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-scripts\") pod \"5d3320db-5949-4f67-8c9b-b709e2e7e542\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.622543 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-scripts" (OuterVolumeSpecName: "scripts") pod "5d3320db-5949-4f67-8c9b-b709e2e7e542" (UID: "5d3320db-5949-4f67-8c9b-b709e2e7e542"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.698082 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-swiftconf\") pod \"5d3320db-5949-4f67-8c9b-b709e2e7e542\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.698197 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-ring-data-devices\") pod \"5d3320db-5949-4f67-8c9b-b709e2e7e542\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.698263 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-dispersionconf\") pod \"5d3320db-5949-4f67-8c9b-b709e2e7e542\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.698305 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w282\" (UniqueName: \"kubernetes.io/projected/5d3320db-5949-4f67-8c9b-b709e2e7e542-kube-api-access-6w282\") pod \"5d3320db-5949-4f67-8c9b-b709e2e7e542\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.698364 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d3320db-5949-4f67-8c9b-b709e2e7e542-etc-swift\") pod \"5d3320db-5949-4f67-8c9b-b709e2e7e542\" (UID: \"5d3320db-5949-4f67-8c9b-b709e2e7e542\") " Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.698956 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.699526 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5d3320db-5949-4f67-8c9b-b709e2e7e542" (UID: "5d3320db-5949-4f67-8c9b-b709e2e7e542"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.699791 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d3320db-5949-4f67-8c9b-b709e2e7e542-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5d3320db-5949-4f67-8c9b-b709e2e7e542" (UID: "5d3320db-5949-4f67-8c9b-b709e2e7e542"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.703001 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d3320db-5949-4f67-8c9b-b709e2e7e542-kube-api-access-6w282" (OuterVolumeSpecName: "kube-api-access-6w282") pod "5d3320db-5949-4f67-8c9b-b709e2e7e542" (UID: "5d3320db-5949-4f67-8c9b-b709e2e7e542"). InnerVolumeSpecName "kube-api-access-6w282". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.722390 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5d3320db-5949-4f67-8c9b-b709e2e7e542" (UID: "5d3320db-5949-4f67-8c9b-b709e2e7e542"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.723286 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5d3320db-5949-4f67-8c9b-b709e2e7e542" (UID: "5d3320db-5949-4f67-8c9b-b709e2e7e542"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.799957 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.800274 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w282\" (UniqueName: \"kubernetes.io/projected/5d3320db-5949-4f67-8c9b-b709e2e7e542-kube-api-access-6w282\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.800343 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5d3320db-5949-4f67-8c9b-b709e2e7e542-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.800405 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5d3320db-5949-4f67-8c9b-b709e2e7e542-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:51 crc kubenswrapper[4692]: I0309 09:56:51.800458 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5d3320db-5949-4f67-8c9b-b709e2e7e542-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.079864 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d3320db-5949-4f67-8c9b-b709e2e7e542" path="/var/lib/kubelet/pods/5d3320db-5949-4f67-8c9b-b709e2e7e542/volumes" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.216438 4692 scope.go:117] "RemoveContainer" containerID="4d7b7a3ddcaea2c88415b27c6976b57b0ca39cda081ce8e68d3ff0496996b830" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.216495 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ms8s8" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.690335 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz"] Mar 09 09:56:52 crc kubenswrapper[4692]: E0309 09:56:52.690978 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d3320db-5949-4f67-8c9b-b709e2e7e542" containerName="swift-ring-rebalance" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.690995 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d3320db-5949-4f67-8c9b-b709e2e7e542" containerName="swift-ring-rebalance" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.691181 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d3320db-5949-4f67-8c9b-b709e2e7e542" containerName="swift-ring-rebalance" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.691754 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.694878 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.695448 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.706443 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz"] Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.817058 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-scripts\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.817182 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/283fc6c9-8646-4336-a88c-19a2e85e09ea-etc-swift\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.817219 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg4bw\" (UniqueName: \"kubernetes.io/projected/283fc6c9-8646-4336-a88c-19a2e85e09ea-kube-api-access-mg4bw\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.817240 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-dispersionconf\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.817279 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-ring-data-devices\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.817303 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-swiftconf\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.918966 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/283fc6c9-8646-4336-a88c-19a2e85e09ea-etc-swift\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.919037 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg4bw\" (UniqueName: \"kubernetes.io/projected/283fc6c9-8646-4336-a88c-19a2e85e09ea-kube-api-access-mg4bw\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.919063 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-dispersionconf\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.919114 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-ring-data-devices\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.919150 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-swiftconf\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.919209 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-scripts\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.919692 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/283fc6c9-8646-4336-a88c-19a2e85e09ea-etc-swift\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.920012 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-scripts\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.920035 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-ring-data-devices\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.924334 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-dispersionconf\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.924596 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-swiftconf\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:52 crc kubenswrapper[4692]: I0309 09:56:52.935508 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg4bw\" (UniqueName: \"kubernetes.io/projected/283fc6c9-8646-4336-a88c-19a2e85e09ea-kube-api-access-mg4bw\") pod \"swift-ring-rebalance-debug-6l9kz\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:53 crc kubenswrapper[4692]: I0309 09:56:53.065653 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:53 crc kubenswrapper[4692]: I0309 09:56:53.501954 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz"] Mar 09 09:56:54 crc kubenswrapper[4692]: I0309 09:56:54.240328 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" event={"ID":"283fc6c9-8646-4336-a88c-19a2e85e09ea","Type":"ContainerStarted","Data":"ddcede9d7cd3b4f95753400e47bbe72d2e55b9a374e9f290f25d0a7aa7289b38"} Mar 09 09:56:54 crc kubenswrapper[4692]: I0309 09:56:54.240470 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" event={"ID":"283fc6c9-8646-4336-a88c-19a2e85e09ea","Type":"ContainerStarted","Data":"3dd562d25a22a1eb7b54bc1aa4d3eb5601eb9ab9075fdc15ea13256805729972"} Mar 09 09:56:54 crc kubenswrapper[4692]: I0309 09:56:54.274954 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" podStartSLOduration=2.274933901 podStartE2EDuration="2.274933901s" podCreationTimestamp="2026-03-09 09:56:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:56:54.273567913 +0000 UTC m=+2215.098303514" watchObservedRunningTime="2026-03-09 09:56:54.274933901 +0000 UTC m=+2215.099669482" Mar 09 09:56:55 crc kubenswrapper[4692]: I0309 09:56:55.253108 4692 generic.go:334] "Generic (PLEG): container finished" podID="283fc6c9-8646-4336-a88c-19a2e85e09ea" containerID="ddcede9d7cd3b4f95753400e47bbe72d2e55b9a374e9f290f25d0a7aa7289b38" exitCode=0 Mar 09 09:56:55 crc kubenswrapper[4692]: I0309 09:56:55.253188 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" event={"ID":"283fc6c9-8646-4336-a88c-19a2e85e09ea","Type":"ContainerDied","Data":"ddcede9d7cd3b4f95753400e47bbe72d2e55b9a374e9f290f25d0a7aa7289b38"} Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.537518 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.571282 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz"] Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.578967 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz"] Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.680216 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-dispersionconf\") pod \"283fc6c9-8646-4336-a88c-19a2e85e09ea\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.680305 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-ring-data-devices\") pod \"283fc6c9-8646-4336-a88c-19a2e85e09ea\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.680382 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/283fc6c9-8646-4336-a88c-19a2e85e09ea-etc-swift\") pod \"283fc6c9-8646-4336-a88c-19a2e85e09ea\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.680426 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg4bw\" (UniqueName: \"kubernetes.io/projected/283fc6c9-8646-4336-a88c-19a2e85e09ea-kube-api-access-mg4bw\") pod \"283fc6c9-8646-4336-a88c-19a2e85e09ea\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.680467 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-scripts\") pod \"283fc6c9-8646-4336-a88c-19a2e85e09ea\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.680554 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-swiftconf\") pod \"283fc6c9-8646-4336-a88c-19a2e85e09ea\" (UID: \"283fc6c9-8646-4336-a88c-19a2e85e09ea\") " Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.681233 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "283fc6c9-8646-4336-a88c-19a2e85e09ea" (UID: "283fc6c9-8646-4336-a88c-19a2e85e09ea"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.681613 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/283fc6c9-8646-4336-a88c-19a2e85e09ea-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "283fc6c9-8646-4336-a88c-19a2e85e09ea" (UID: "283fc6c9-8646-4336-a88c-19a2e85e09ea"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.689255 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/283fc6c9-8646-4336-a88c-19a2e85e09ea-kube-api-access-mg4bw" (OuterVolumeSpecName: "kube-api-access-mg4bw") pod "283fc6c9-8646-4336-a88c-19a2e85e09ea" (UID: "283fc6c9-8646-4336-a88c-19a2e85e09ea"). InnerVolumeSpecName "kube-api-access-mg4bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.701932 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-scripts" (OuterVolumeSpecName: "scripts") pod "283fc6c9-8646-4336-a88c-19a2e85e09ea" (UID: "283fc6c9-8646-4336-a88c-19a2e85e09ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.705242 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "283fc6c9-8646-4336-a88c-19a2e85e09ea" (UID: "283fc6c9-8646-4336-a88c-19a2e85e09ea"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.712078 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "283fc6c9-8646-4336-a88c-19a2e85e09ea" (UID: "283fc6c9-8646-4336-a88c-19a2e85e09ea"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.782240 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.782290 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/283fc6c9-8646-4336-a88c-19a2e85e09ea-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.782301 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg4bw\" (UniqueName: \"kubernetes.io/projected/283fc6c9-8646-4336-a88c-19a2e85e09ea-kube-api-access-mg4bw\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.782316 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/283fc6c9-8646-4336-a88c-19a2e85e09ea-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.782328 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:56 crc kubenswrapper[4692]: I0309 09:56:56.782337 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/283fc6c9-8646-4336-a88c-19a2e85e09ea-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.282654 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dd562d25a22a1eb7b54bc1aa4d3eb5601eb9ab9075fdc15ea13256805729972" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.282788 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6l9kz" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.729755 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs"] Mar 09 09:56:57 crc kubenswrapper[4692]: E0309 09:56:57.730134 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="283fc6c9-8646-4336-a88c-19a2e85e09ea" containerName="swift-ring-rebalance" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.730150 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="283fc6c9-8646-4336-a88c-19a2e85e09ea" containerName="swift-ring-rebalance" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.730317 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="283fc6c9-8646-4336-a88c-19a2e85e09ea" containerName="swift-ring-rebalance" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.730924 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.733363 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.734589 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.749436 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs"] Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.803556 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-dispersionconf\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.803626 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6bt6\" (UniqueName: \"kubernetes.io/projected/1d088dd9-9d45-4569-b828-610820dfc952-kube-api-access-x6bt6\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.803678 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-scripts\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.803717 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-ring-data-devices\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.803749 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-swiftconf\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.803796 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d088dd9-9d45-4569-b828-610820dfc952-etc-swift\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.904871 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-dispersionconf\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.904935 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6bt6\" (UniqueName: \"kubernetes.io/projected/1d088dd9-9d45-4569-b828-610820dfc952-kube-api-access-x6bt6\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.904965 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-scripts\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.904996 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-ring-data-devices\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.905026 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-swiftconf\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.905058 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d088dd9-9d45-4569-b828-610820dfc952-etc-swift\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.905589 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d088dd9-9d45-4569-b828-610820dfc952-etc-swift\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.906398 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-ring-data-devices\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.906799 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-scripts\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.916801 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-swiftconf\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.917196 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-dispersionconf\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:57 crc kubenswrapper[4692]: I0309 09:56:57.928274 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6bt6\" (UniqueName: \"kubernetes.io/projected/1d088dd9-9d45-4569-b828-610820dfc952-kube-api-access-x6bt6\") pod \"swift-ring-rebalance-debug-zgmqs\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:58 crc kubenswrapper[4692]: I0309 09:56:58.057748 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:56:58 crc kubenswrapper[4692]: I0309 09:56:58.083864 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="283fc6c9-8646-4336-a88c-19a2e85e09ea" path="/var/lib/kubelet/pods/283fc6c9-8646-4336-a88c-19a2e85e09ea/volumes" Mar 09 09:56:58 crc kubenswrapper[4692]: I0309 09:56:58.502019 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs"] Mar 09 09:56:58 crc kubenswrapper[4692]: W0309 09:56:58.509148 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d088dd9_9d45_4569_b828_610820dfc952.slice/crio-9c833a30442b00e7860f7eb5fecef62c4e54c4ea140d23111978de1e022104af WatchSource:0}: Error finding container 9c833a30442b00e7860f7eb5fecef62c4e54c4ea140d23111978de1e022104af: Status 404 returned error can't find the container with id 9c833a30442b00e7860f7eb5fecef62c4e54c4ea140d23111978de1e022104af Mar 09 09:56:59 crc kubenswrapper[4692]: I0309 09:56:59.300519 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" event={"ID":"1d088dd9-9d45-4569-b828-610820dfc952","Type":"ContainerStarted","Data":"c340486dec4140fe4248064cefeacb0343a945c850cad805cc96aaff8db5e408"} Mar 09 09:56:59 crc kubenswrapper[4692]: I0309 09:56:59.300569 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" event={"ID":"1d088dd9-9d45-4569-b828-610820dfc952","Type":"ContainerStarted","Data":"9c833a30442b00e7860f7eb5fecef62c4e54c4ea140d23111978de1e022104af"} Mar 09 09:57:00 crc kubenswrapper[4692]: I0309 09:57:00.315102 4692 generic.go:334] "Generic (PLEG): container finished" podID="1d088dd9-9d45-4569-b828-610820dfc952" containerID="c340486dec4140fe4248064cefeacb0343a945c850cad805cc96aaff8db5e408" exitCode=0 Mar 09 09:57:00 crc kubenswrapper[4692]: I0309 09:57:00.315195 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" event={"ID":"1d088dd9-9d45-4569-b828-610820dfc952","Type":"ContainerDied","Data":"c340486dec4140fe4248064cefeacb0343a945c850cad805cc96aaff8db5e408"} Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.659462 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.674892 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-dispersionconf\") pod \"1d088dd9-9d45-4569-b828-610820dfc952\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.675005 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-ring-data-devices\") pod \"1d088dd9-9d45-4569-b828-610820dfc952\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.675033 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-swiftconf\") pod \"1d088dd9-9d45-4569-b828-610820dfc952\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.675078 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d088dd9-9d45-4569-b828-610820dfc952-etc-swift\") pod \"1d088dd9-9d45-4569-b828-610820dfc952\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.675146 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-scripts\") pod \"1d088dd9-9d45-4569-b828-610820dfc952\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.675211 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6bt6\" (UniqueName: \"kubernetes.io/projected/1d088dd9-9d45-4569-b828-610820dfc952-kube-api-access-x6bt6\") pod \"1d088dd9-9d45-4569-b828-610820dfc952\" (UID: \"1d088dd9-9d45-4569-b828-610820dfc952\") " Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.675693 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "1d088dd9-9d45-4569-b828-610820dfc952" (UID: "1d088dd9-9d45-4569-b828-610820dfc952"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.675923 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d088dd9-9d45-4569-b828-610820dfc952-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "1d088dd9-9d45-4569-b828-610820dfc952" (UID: "1d088dd9-9d45-4569-b828-610820dfc952"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.682112 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d088dd9-9d45-4569-b828-610820dfc952-kube-api-access-x6bt6" (OuterVolumeSpecName: "kube-api-access-x6bt6") pod "1d088dd9-9d45-4569-b828-610820dfc952" (UID: "1d088dd9-9d45-4569-b828-610820dfc952"). InnerVolumeSpecName "kube-api-access-x6bt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.696229 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs"] Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.698445 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-scripts" (OuterVolumeSpecName: "scripts") pod "1d088dd9-9d45-4569-b828-610820dfc952" (UID: "1d088dd9-9d45-4569-b828-610820dfc952"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.701923 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs"] Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.714037 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "1d088dd9-9d45-4569-b828-610820dfc952" (UID: "1d088dd9-9d45-4569-b828-610820dfc952"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.714283 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "1d088dd9-9d45-4569-b828-610820dfc952" (UID: "1d088dd9-9d45-4569-b828-610820dfc952"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.776833 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.776882 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.776896 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/1d088dd9-9d45-4569-b828-610820dfc952-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.776906 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/1d088dd9-9d45-4569-b828-610820dfc952-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.776918 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d088dd9-9d45-4569-b828-610820dfc952-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:01 crc kubenswrapper[4692]: I0309 09:57:01.776928 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6bt6\" (UniqueName: \"kubernetes.io/projected/1d088dd9-9d45-4569-b828-610820dfc952-kube-api-access-x6bt6\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.083998 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d088dd9-9d45-4569-b828-610820dfc952" path="/var/lib/kubelet/pods/1d088dd9-9d45-4569-b828-610820dfc952/volumes" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.338864 4692 scope.go:117] "RemoveContainer" containerID="c340486dec4140fe4248064cefeacb0343a945c850cad805cc96aaff8db5e408" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.338921 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zgmqs" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.848984 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm"] Mar 09 09:57:02 crc kubenswrapper[4692]: E0309 09:57:02.849390 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d088dd9-9d45-4569-b828-610820dfc952" containerName="swift-ring-rebalance" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.849407 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d088dd9-9d45-4569-b828-610820dfc952" containerName="swift-ring-rebalance" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.849596 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d088dd9-9d45-4569-b828-610820dfc952" containerName="swift-ring-rebalance" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.850267 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.854108 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.854646 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.861729 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm"] Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.995611 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-ring-data-devices\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.995779 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-scripts\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.995860 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-dispersionconf\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.995984 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a24e329-0db1-4a04-b745-31a48e6f5e1c-etc-swift\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.996077 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-swiftconf\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:02 crc kubenswrapper[4692]: I0309 09:57:02.996275 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdsjl\" (UniqueName: \"kubernetes.io/projected/3a24e329-0db1-4a04-b745-31a48e6f5e1c-kube-api-access-mdsjl\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.097678 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdsjl\" (UniqueName: \"kubernetes.io/projected/3a24e329-0db1-4a04-b745-31a48e6f5e1c-kube-api-access-mdsjl\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.097759 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-ring-data-devices\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.097800 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-scripts\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.097814 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-dispersionconf\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.097851 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a24e329-0db1-4a04-b745-31a48e6f5e1c-etc-swift\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.097899 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-swiftconf\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.098890 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-scripts\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.099318 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-ring-data-devices\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.099859 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a24e329-0db1-4a04-b745-31a48e6f5e1c-etc-swift\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.104017 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-swiftconf\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.104090 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-dispersionconf\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.116681 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdsjl\" (UniqueName: \"kubernetes.io/projected/3a24e329-0db1-4a04-b745-31a48e6f5e1c-kube-api-access-mdsjl\") pod \"swift-ring-rebalance-debug-gz9lm\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.180684 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:03 crc kubenswrapper[4692]: I0309 09:57:03.622293 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm"] Mar 09 09:57:03 crc kubenswrapper[4692]: W0309 09:57:03.627299 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a24e329_0db1_4a04_b745_31a48e6f5e1c.slice/crio-5adcb1a0e4c59a08289d4bd069092690f0b48cef5a64e82439196133dbac88d0 WatchSource:0}: Error finding container 5adcb1a0e4c59a08289d4bd069092690f0b48cef5a64e82439196133dbac88d0: Status 404 returned error can't find the container with id 5adcb1a0e4c59a08289d4bd069092690f0b48cef5a64e82439196133dbac88d0 Mar 09 09:57:04 crc kubenswrapper[4692]: I0309 09:57:04.366412 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" event={"ID":"3a24e329-0db1-4a04-b745-31a48e6f5e1c","Type":"ContainerStarted","Data":"787409a72bfb1e6a5f6c151022b0437e7081acfbac56da9635bbd21aa799c6bb"} Mar 09 09:57:04 crc kubenswrapper[4692]: I0309 09:57:04.366942 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" event={"ID":"3a24e329-0db1-4a04-b745-31a48e6f5e1c","Type":"ContainerStarted","Data":"5adcb1a0e4c59a08289d4bd069092690f0b48cef5a64e82439196133dbac88d0"} Mar 09 09:57:04 crc kubenswrapper[4692]: I0309 09:57:04.406684 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" podStartSLOduration=2.4066475929999998 podStartE2EDuration="2.406647593s" podCreationTimestamp="2026-03-09 09:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:04.390899978 +0000 UTC m=+2225.215635569" watchObservedRunningTime="2026-03-09 09:57:04.406647593 +0000 UTC m=+2225.231383214" Mar 09 09:57:04 crc kubenswrapper[4692]: I0309 09:57:04.433250 4692 scope.go:117] "RemoveContainer" containerID="005da3fb4a7d971c028d2a59b2a17104c914537c845235e63f1413e6f7bc3b8d" Mar 09 09:57:05 crc kubenswrapper[4692]: I0309 09:57:05.382407 4692 generic.go:334] "Generic (PLEG): container finished" podID="3a24e329-0db1-4a04-b745-31a48e6f5e1c" containerID="787409a72bfb1e6a5f6c151022b0437e7081acfbac56da9635bbd21aa799c6bb" exitCode=0 Mar 09 09:57:05 crc kubenswrapper[4692]: I0309 09:57:05.382494 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" event={"ID":"3a24e329-0db1-4a04-b745-31a48e6f5e1c","Type":"ContainerDied","Data":"787409a72bfb1e6a5f6c151022b0437e7081acfbac56da9635bbd21aa799c6bb"} Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.704067 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.745155 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm"] Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.758238 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm"] Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.866221 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-scripts\") pod \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.866445 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a24e329-0db1-4a04-b745-31a48e6f5e1c-etc-swift\") pod \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.866713 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-ring-data-devices\") pod \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.866850 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-dispersionconf\") pod \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.866887 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdsjl\" (UniqueName: \"kubernetes.io/projected/3a24e329-0db1-4a04-b745-31a48e6f5e1c-kube-api-access-mdsjl\") pod \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.866931 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-swiftconf\") pod \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\" (UID: \"3a24e329-0db1-4a04-b745-31a48e6f5e1c\") " Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.867860 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3a24e329-0db1-4a04-b745-31a48e6f5e1c" (UID: "3a24e329-0db1-4a04-b745-31a48e6f5e1c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.868133 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a24e329-0db1-4a04-b745-31a48e6f5e1c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3a24e329-0db1-4a04-b745-31a48e6f5e1c" (UID: "3a24e329-0db1-4a04-b745-31a48e6f5e1c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.873283 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a24e329-0db1-4a04-b745-31a48e6f5e1c-kube-api-access-mdsjl" (OuterVolumeSpecName: "kube-api-access-mdsjl") pod "3a24e329-0db1-4a04-b745-31a48e6f5e1c" (UID: "3a24e329-0db1-4a04-b745-31a48e6f5e1c"). InnerVolumeSpecName "kube-api-access-mdsjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.893042 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3a24e329-0db1-4a04-b745-31a48e6f5e1c" (UID: "3a24e329-0db1-4a04-b745-31a48e6f5e1c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.893507 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-scripts" (OuterVolumeSpecName: "scripts") pod "3a24e329-0db1-4a04-b745-31a48e6f5e1c" (UID: "3a24e329-0db1-4a04-b745-31a48e6f5e1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.895496 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3a24e329-0db1-4a04-b745-31a48e6f5e1c" (UID: "3a24e329-0db1-4a04-b745-31a48e6f5e1c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.968937 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.968974 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdsjl\" (UniqueName: \"kubernetes.io/projected/3a24e329-0db1-4a04-b745-31a48e6f5e1c-kube-api-access-mdsjl\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.968993 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3a24e329-0db1-4a04-b745-31a48e6f5e1c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.969002 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.969013 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3a24e329-0db1-4a04-b745-31a48e6f5e1c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:06 crc kubenswrapper[4692]: I0309 09:57:06.969021 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3a24e329-0db1-4a04-b745-31a48e6f5e1c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.410459 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5adcb1a0e4c59a08289d4bd069092690f0b48cef5a64e82439196133dbac88d0" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.410577 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gz9lm" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.911813 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qxthn"] Mar 09 09:57:07 crc kubenswrapper[4692]: E0309 09:57:07.912889 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a24e329-0db1-4a04-b745-31a48e6f5e1c" containerName="swift-ring-rebalance" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.912913 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a24e329-0db1-4a04-b745-31a48e6f5e1c" containerName="swift-ring-rebalance" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.913142 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a24e329-0db1-4a04-b745-31a48e6f5e1c" containerName="swift-ring-rebalance" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.914844 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.917419 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.917915 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:07 crc kubenswrapper[4692]: I0309 09:57:07.927020 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qxthn"] Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.081384 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a24e329-0db1-4a04-b745-31a48e6f5e1c" path="/var/lib/kubelet/pods/3a24e329-0db1-4a04-b745-31a48e6f5e1c/volumes" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.088952 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2798d153-9e7c-4044-b5f0-0477b868f010-etc-swift\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.089693 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-scripts\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.089876 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-dispersionconf\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.090038 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-swiftconf\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.090155 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgm7j\" (UniqueName: \"kubernetes.io/projected/2798d153-9e7c-4044-b5f0-0477b868f010-kube-api-access-pgm7j\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.090406 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-ring-data-devices\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.192748 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2798d153-9e7c-4044-b5f0-0477b868f010-etc-swift\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.192854 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-scripts\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.192879 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-dispersionconf\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.192900 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-swiftconf\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.192920 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgm7j\" (UniqueName: \"kubernetes.io/projected/2798d153-9e7c-4044-b5f0-0477b868f010-kube-api-access-pgm7j\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.192965 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-ring-data-devices\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.193509 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2798d153-9e7c-4044-b5f0-0477b868f010-etc-swift\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.193797 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-ring-data-devices\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.194327 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-scripts\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.198400 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-dispersionconf\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.198477 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-swiftconf\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.210895 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgm7j\" (UniqueName: \"kubernetes.io/projected/2798d153-9e7c-4044-b5f0-0477b868f010-kube-api-access-pgm7j\") pod \"swift-ring-rebalance-debug-qxthn\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.237070 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:08 crc kubenswrapper[4692]: I0309 09:57:08.661765 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qxthn"] Mar 09 09:57:08 crc kubenswrapper[4692]: W0309 09:57:08.667717 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2798d153_9e7c_4044_b5f0_0477b868f010.slice/crio-6f9b3d30698a6afc9f322c62162f24537bb312d3483247f54d29faec92e69913 WatchSource:0}: Error finding container 6f9b3d30698a6afc9f322c62162f24537bb312d3483247f54d29faec92e69913: Status 404 returned error can't find the container with id 6f9b3d30698a6afc9f322c62162f24537bb312d3483247f54d29faec92e69913 Mar 09 09:57:09 crc kubenswrapper[4692]: I0309 09:57:09.433623 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" event={"ID":"2798d153-9e7c-4044-b5f0-0477b868f010","Type":"ContainerStarted","Data":"c0d6c2246f07b6f7978c796868390c4d45b42a025c60cc86d46564b97fab1776"} Mar 09 09:57:09 crc kubenswrapper[4692]: I0309 09:57:09.434027 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" event={"ID":"2798d153-9e7c-4044-b5f0-0477b868f010","Type":"ContainerStarted","Data":"6f9b3d30698a6afc9f322c62162f24537bb312d3483247f54d29faec92e69913"} Mar 09 09:57:09 crc kubenswrapper[4692]: I0309 09:57:09.460840 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" podStartSLOduration=2.460822567 podStartE2EDuration="2.460822567s" podCreationTimestamp="2026-03-09 09:57:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:09.458381718 +0000 UTC m=+2230.283117369" watchObservedRunningTime="2026-03-09 09:57:09.460822567 +0000 UTC m=+2230.285558148" Mar 09 09:57:10 crc kubenswrapper[4692]: I0309 09:57:10.443884 4692 generic.go:334] "Generic (PLEG): container finished" podID="2798d153-9e7c-4044-b5f0-0477b868f010" containerID="c0d6c2246f07b6f7978c796868390c4d45b42a025c60cc86d46564b97fab1776" exitCode=0 Mar 09 09:57:10 crc kubenswrapper[4692]: I0309 09:57:10.444010 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" event={"ID":"2798d153-9e7c-4044-b5f0-0477b868f010","Type":"ContainerDied","Data":"c0d6c2246f07b6f7978c796868390c4d45b42a025c60cc86d46564b97fab1776"} Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.709401 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.772437 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qxthn"] Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.777817 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qxthn"] Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.865920 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-ring-data-devices\") pod \"2798d153-9e7c-4044-b5f0-0477b868f010\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.866381 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgm7j\" (UniqueName: \"kubernetes.io/projected/2798d153-9e7c-4044-b5f0-0477b868f010-kube-api-access-pgm7j\") pod \"2798d153-9e7c-4044-b5f0-0477b868f010\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.866491 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2798d153-9e7c-4044-b5f0-0477b868f010-etc-swift\") pod \"2798d153-9e7c-4044-b5f0-0477b868f010\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.867329 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-scripts\") pod \"2798d153-9e7c-4044-b5f0-0477b868f010\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.867420 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-dispersionconf\") pod \"2798d153-9e7c-4044-b5f0-0477b868f010\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.867961 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-swiftconf\") pod \"2798d153-9e7c-4044-b5f0-0477b868f010\" (UID: \"2798d153-9e7c-4044-b5f0-0477b868f010\") " Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.866929 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2798d153-9e7c-4044-b5f0-0477b868f010" (UID: "2798d153-9e7c-4044-b5f0-0477b868f010"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.867260 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2798d153-9e7c-4044-b5f0-0477b868f010-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2798d153-9e7c-4044-b5f0-0477b868f010" (UID: "2798d153-9e7c-4044-b5f0-0477b868f010"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.868533 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.868605 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2798d153-9e7c-4044-b5f0-0477b868f010-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.873811 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2798d153-9e7c-4044-b5f0-0477b868f010-kube-api-access-pgm7j" (OuterVolumeSpecName: "kube-api-access-pgm7j") pod "2798d153-9e7c-4044-b5f0-0477b868f010" (UID: "2798d153-9e7c-4044-b5f0-0477b868f010"). InnerVolumeSpecName "kube-api-access-pgm7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.893731 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-scripts" (OuterVolumeSpecName: "scripts") pod "2798d153-9e7c-4044-b5f0-0477b868f010" (UID: "2798d153-9e7c-4044-b5f0-0477b868f010"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.898128 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2798d153-9e7c-4044-b5f0-0477b868f010" (UID: "2798d153-9e7c-4044-b5f0-0477b868f010"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.908937 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2798d153-9e7c-4044-b5f0-0477b868f010" (UID: "2798d153-9e7c-4044-b5f0-0477b868f010"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.969536 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgm7j\" (UniqueName: \"kubernetes.io/projected/2798d153-9e7c-4044-b5f0-0477b868f010-kube-api-access-pgm7j\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.969577 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2798d153-9e7c-4044-b5f0-0477b868f010-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.969591 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:11 crc kubenswrapper[4692]: I0309 09:57:11.969600 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2798d153-9e7c-4044-b5f0-0477b868f010-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.080609 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2798d153-9e7c-4044-b5f0-0477b868f010" path="/var/lib/kubelet/pods/2798d153-9e7c-4044-b5f0-0477b868f010/volumes" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.470599 4692 scope.go:117] "RemoveContainer" containerID="c0d6c2246f07b6f7978c796868390c4d45b42a025c60cc86d46564b97fab1776" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.470666 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qxthn" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.894661 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7"] Mar 09 09:57:12 crc kubenswrapper[4692]: E0309 09:57:12.895724 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2798d153-9e7c-4044-b5f0-0477b868f010" containerName="swift-ring-rebalance" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.895750 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2798d153-9e7c-4044-b5f0-0477b868f010" containerName="swift-ring-rebalance" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.896039 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2798d153-9e7c-4044-b5f0-0477b868f010" containerName="swift-ring-rebalance" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.897056 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.901189 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.901563 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.908566 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7"] Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.986054 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-dispersionconf\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.986411 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qntch\" (UniqueName: \"kubernetes.io/projected/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-kube-api-access-qntch\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.986506 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-scripts\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.986569 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-swiftconf\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.986739 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-etc-swift\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:12 crc kubenswrapper[4692]: I0309 09:57:12.986857 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-ring-data-devices\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.088349 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-ring-data-devices\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.088476 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-dispersionconf\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.088529 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qntch\" (UniqueName: \"kubernetes.io/projected/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-kube-api-access-qntch\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.088556 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-scripts\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.088579 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-swiftconf\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.088603 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-etc-swift\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.090207 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-etc-swift\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.091022 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-ring-data-devices\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.091295 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-scripts\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.095430 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-dispersionconf\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.104120 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-swiftconf\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.117471 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qntch\" (UniqueName: \"kubernetes.io/projected/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-kube-api-access-qntch\") pod \"swift-ring-rebalance-debug-sjlj7\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.219778 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:13 crc kubenswrapper[4692]: I0309 09:57:13.719823 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7"] Mar 09 09:57:14 crc kubenswrapper[4692]: I0309 09:57:14.501743 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" event={"ID":"b7dddffc-965c-453c-bde7-8eb25b5d6e2d","Type":"ContainerStarted","Data":"d6cc48a65007fc5e3a81661d7c4042baff65e5b33a69f27bcde40c180d1cea5e"} Mar 09 09:57:14 crc kubenswrapper[4692]: I0309 09:57:14.502123 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" event={"ID":"b7dddffc-965c-453c-bde7-8eb25b5d6e2d","Type":"ContainerStarted","Data":"d3c85b2b9d2922468dd2656e88bc26415b74476f1b77ffbf387b345c8a4d044f"} Mar 09 09:57:14 crc kubenswrapper[4692]: I0309 09:57:14.535951 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" podStartSLOduration=2.535923204 podStartE2EDuration="2.535923204s" podCreationTimestamp="2026-03-09 09:57:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:14.5294174 +0000 UTC m=+2235.354153021" watchObservedRunningTime="2026-03-09 09:57:14.535923204 +0000 UTC m=+2235.360658785" Mar 09 09:57:15 crc kubenswrapper[4692]: I0309 09:57:15.511084 4692 generic.go:334] "Generic (PLEG): container finished" podID="b7dddffc-965c-453c-bde7-8eb25b5d6e2d" containerID="d6cc48a65007fc5e3a81661d7c4042baff65e5b33a69f27bcde40c180d1cea5e" exitCode=0 Mar 09 09:57:15 crc kubenswrapper[4692]: I0309 09:57:15.511130 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" event={"ID":"b7dddffc-965c-453c-bde7-8eb25b5d6e2d","Type":"ContainerDied","Data":"d6cc48a65007fc5e3a81661d7c4042baff65e5b33a69f27bcde40c180d1cea5e"} Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.813718 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.855461 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7"] Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.862869 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7"] Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.957884 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-swiftconf\") pod \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.957944 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-etc-swift\") pod \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.957987 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qntch\" (UniqueName: \"kubernetes.io/projected/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-kube-api-access-qntch\") pod \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.958014 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-scripts\") pod \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.958123 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-dispersionconf\") pod \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.958196 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-ring-data-devices\") pod \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\" (UID: \"b7dddffc-965c-453c-bde7-8eb25b5d6e2d\") " Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.959026 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b7dddffc-965c-453c-bde7-8eb25b5d6e2d" (UID: "b7dddffc-965c-453c-bde7-8eb25b5d6e2d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.959269 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b7dddffc-965c-453c-bde7-8eb25b5d6e2d" (UID: "b7dddffc-965c-453c-bde7-8eb25b5d6e2d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.965386 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-kube-api-access-qntch" (OuterVolumeSpecName: "kube-api-access-qntch") pod "b7dddffc-965c-453c-bde7-8eb25b5d6e2d" (UID: "b7dddffc-965c-453c-bde7-8eb25b5d6e2d"). InnerVolumeSpecName "kube-api-access-qntch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.987456 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b7dddffc-965c-453c-bde7-8eb25b5d6e2d" (UID: "b7dddffc-965c-453c-bde7-8eb25b5d6e2d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.993645 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-scripts" (OuterVolumeSpecName: "scripts") pod "b7dddffc-965c-453c-bde7-8eb25b5d6e2d" (UID: "b7dddffc-965c-453c-bde7-8eb25b5d6e2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:16 crc kubenswrapper[4692]: I0309 09:57:16.994787 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b7dddffc-965c-453c-bde7-8eb25b5d6e2d" (UID: "b7dddffc-965c-453c-bde7-8eb25b5d6e2d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.060148 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.060226 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.060237 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.060247 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.060258 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qntch\" (UniqueName: \"kubernetes.io/projected/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-kube-api-access-qntch\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.060270 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b7dddffc-965c-453c-bde7-8eb25b5d6e2d-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.550454 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3c85b2b9d2922468dd2656e88bc26415b74476f1b77ffbf387b345c8a4d044f" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.550496 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-sjlj7" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.624329 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.624443 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.624487 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.625217 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 09:57:17 crc kubenswrapper[4692]: I0309 09:57:17.625280 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" gracePeriod=600 Mar 09 09:57:17 crc kubenswrapper[4692]: E0309 09:57:17.759015 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.029777 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j9q49"] Mar 09 09:57:18 crc kubenswrapper[4692]: E0309 09:57:18.030097 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7dddffc-965c-453c-bde7-8eb25b5d6e2d" containerName="swift-ring-rebalance" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.030109 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7dddffc-965c-453c-bde7-8eb25b5d6e2d" containerName="swift-ring-rebalance" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.030296 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7dddffc-965c-453c-bde7-8eb25b5d6e2d" containerName="swift-ring-rebalance" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.030725 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.034018 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.034271 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.044929 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j9q49"] Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.094885 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7dddffc-965c-453c-bde7-8eb25b5d6e2d" path="/var/lib/kubelet/pods/b7dddffc-965c-453c-bde7-8eb25b5d6e2d/volumes" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.179576 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-dispersionconf\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.179640 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-etc-swift\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.179668 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-ring-data-devices\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.179715 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clzxq\" (UniqueName: \"kubernetes.io/projected/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-kube-api-access-clzxq\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.179741 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-scripts\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.179793 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-swiftconf\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.280951 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-dispersionconf\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.281267 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-etc-swift\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.281294 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-ring-data-devices\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.281326 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clzxq\" (UniqueName: \"kubernetes.io/projected/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-kube-api-access-clzxq\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.281356 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-scripts\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.281426 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-swiftconf\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.282228 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-etc-swift\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.282288 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-ring-data-devices\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.282758 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-scripts\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.286184 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-swiftconf\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.289629 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-dispersionconf\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.298950 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clzxq\" (UniqueName: \"kubernetes.io/projected/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-kube-api-access-clzxq\") pod \"swift-ring-rebalance-debug-j9q49\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.354865 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.560245 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" exitCode=0 Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.560260 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff"} Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.560324 4692 scope.go:117] "RemoveContainer" containerID="cfb6f9265bcdee5d06e08425c1dd3a025aa1540a263740462908a83a3191a85f" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.561358 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:57:18 crc kubenswrapper[4692]: E0309 09:57:18.561638 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:57:18 crc kubenswrapper[4692]: I0309 09:57:18.749358 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j9q49"] Mar 09 09:57:18 crc kubenswrapper[4692]: W0309 09:57:18.750581 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dd767f2_ac32_456d_b9ac_fe0ee4115d16.slice/crio-faf15e195ce851143d616eb3ebd912532ee359320643c130022222defe4deac0 WatchSource:0}: Error finding container faf15e195ce851143d616eb3ebd912532ee359320643c130022222defe4deac0: Status 404 returned error can't find the container with id faf15e195ce851143d616eb3ebd912532ee359320643c130022222defe4deac0 Mar 09 09:57:19 crc kubenswrapper[4692]: I0309 09:57:19.573080 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" event={"ID":"3dd767f2-ac32-456d-b9ac-fe0ee4115d16","Type":"ContainerStarted","Data":"11d454baf69231055f065d2625a39865ee14cff70257a71b264e8ee440034bbb"} Mar 09 09:57:19 crc kubenswrapper[4692]: I0309 09:57:19.573433 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" event={"ID":"3dd767f2-ac32-456d-b9ac-fe0ee4115d16","Type":"ContainerStarted","Data":"faf15e195ce851143d616eb3ebd912532ee359320643c130022222defe4deac0"} Mar 09 09:57:20 crc kubenswrapper[4692]: I0309 09:57:20.585458 4692 generic.go:334] "Generic (PLEG): container finished" podID="3dd767f2-ac32-456d-b9ac-fe0ee4115d16" containerID="11d454baf69231055f065d2625a39865ee14cff70257a71b264e8ee440034bbb" exitCode=0 Mar 09 09:57:20 crc kubenswrapper[4692]: I0309 09:57:20.585502 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" event={"ID":"3dd767f2-ac32-456d-b9ac-fe0ee4115d16","Type":"ContainerDied","Data":"11d454baf69231055f065d2625a39865ee14cff70257a71b264e8ee440034bbb"} Mar 09 09:57:21 crc kubenswrapper[4692]: I0309 09:57:21.892960 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:21 crc kubenswrapper[4692]: I0309 09:57:21.919682 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j9q49"] Mar 09 09:57:21 crc kubenswrapper[4692]: I0309 09:57:21.925731 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j9q49"] Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.030915 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-etc-swift\") pod \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.031043 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-scripts\") pod \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.031084 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clzxq\" (UniqueName: \"kubernetes.io/projected/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-kube-api-access-clzxq\") pod \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.031111 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-dispersionconf\") pod \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.031182 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-swiftconf\") pod \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.031211 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-ring-data-devices\") pod \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\" (UID: \"3dd767f2-ac32-456d-b9ac-fe0ee4115d16\") " Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.031923 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3dd767f2-ac32-456d-b9ac-fe0ee4115d16" (UID: "3dd767f2-ac32-456d-b9ac-fe0ee4115d16"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.031957 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3dd767f2-ac32-456d-b9ac-fe0ee4115d16" (UID: "3dd767f2-ac32-456d-b9ac-fe0ee4115d16"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.035882 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-kube-api-access-clzxq" (OuterVolumeSpecName: "kube-api-access-clzxq") pod "3dd767f2-ac32-456d-b9ac-fe0ee4115d16" (UID: "3dd767f2-ac32-456d-b9ac-fe0ee4115d16"). InnerVolumeSpecName "kube-api-access-clzxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.053553 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3dd767f2-ac32-456d-b9ac-fe0ee4115d16" (UID: "3dd767f2-ac32-456d-b9ac-fe0ee4115d16"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.054476 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3dd767f2-ac32-456d-b9ac-fe0ee4115d16" (UID: "3dd767f2-ac32-456d-b9ac-fe0ee4115d16"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.064983 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-scripts" (OuterVolumeSpecName: "scripts") pod "3dd767f2-ac32-456d-b9ac-fe0ee4115d16" (UID: "3dd767f2-ac32-456d-b9ac-fe0ee4115d16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.080655 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dd767f2-ac32-456d-b9ac-fe0ee4115d16" path="/var/lib/kubelet/pods/3dd767f2-ac32-456d-b9ac-fe0ee4115d16/volumes" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.133314 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.133346 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.133358 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.133369 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.133379 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clzxq\" (UniqueName: \"kubernetes.io/projected/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-kube-api-access-clzxq\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.133386 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3dd767f2-ac32-456d-b9ac-fe0ee4115d16-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.604575 4692 scope.go:117] "RemoveContainer" containerID="11d454baf69231055f065d2625a39865ee14cff70257a71b264e8ee440034bbb" Mar 09 09:57:22 crc kubenswrapper[4692]: I0309 09:57:22.605346 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j9q49" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.073431 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-69vn9"] Mar 09 09:57:23 crc kubenswrapper[4692]: E0309 09:57:23.075292 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd767f2-ac32-456d-b9ac-fe0ee4115d16" containerName="swift-ring-rebalance" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.075345 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd767f2-ac32-456d-b9ac-fe0ee4115d16" containerName="swift-ring-rebalance" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.076281 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd767f2-ac32-456d-b9ac-fe0ee4115d16" containerName="swift-ring-rebalance" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.078188 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.087315 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.088175 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.093430 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-69vn9"] Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.250889 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-ring-data-devices\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.250948 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68efd13e-278f-4166-bf74-64e0cf2b5b9a-etc-swift\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.250993 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-swiftconf\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.251010 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-dispersionconf\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.251046 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-scripts\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.251084 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvk7r\" (UniqueName: \"kubernetes.io/projected/68efd13e-278f-4166-bf74-64e0cf2b5b9a-kube-api-access-gvk7r\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.352818 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-swiftconf\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.353099 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-dispersionconf\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.353121 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-scripts\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.353179 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvk7r\" (UniqueName: \"kubernetes.io/projected/68efd13e-278f-4166-bf74-64e0cf2b5b9a-kube-api-access-gvk7r\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.353261 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-ring-data-devices\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.353290 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68efd13e-278f-4166-bf74-64e0cf2b5b9a-etc-swift\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.353786 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68efd13e-278f-4166-bf74-64e0cf2b5b9a-etc-swift\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.354371 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-scripts\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.354406 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-ring-data-devices\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.358487 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-swiftconf\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.359487 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-dispersionconf\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.373500 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvk7r\" (UniqueName: \"kubernetes.io/projected/68efd13e-278f-4166-bf74-64e0cf2b5b9a-kube-api-access-gvk7r\") pod \"swift-ring-rebalance-debug-69vn9\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.400408 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:23 crc kubenswrapper[4692]: I0309 09:57:23.872872 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-69vn9"] Mar 09 09:57:24 crc kubenswrapper[4692]: I0309 09:57:24.622709 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" event={"ID":"68efd13e-278f-4166-bf74-64e0cf2b5b9a","Type":"ContainerStarted","Data":"9e74e40ad1920b57ede3cf39989dae9ba4843dc6b52db3800f65d3b52cfe29b4"} Mar 09 09:57:24 crc kubenswrapper[4692]: I0309 09:57:24.622776 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" event={"ID":"68efd13e-278f-4166-bf74-64e0cf2b5b9a","Type":"ContainerStarted","Data":"9316cbb361e3796954fb1e2f4c2d09085a99ad1e120a15ae48f14e27f4d7f086"} Mar 09 09:57:24 crc kubenswrapper[4692]: I0309 09:57:24.646609 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" podStartSLOduration=1.6465851900000001 podStartE2EDuration="1.64658519s" podCreationTimestamp="2026-03-09 09:57:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:24.638023608 +0000 UTC m=+2245.462759239" watchObservedRunningTime="2026-03-09 09:57:24.64658519 +0000 UTC m=+2245.471320811" Mar 09 09:57:25 crc kubenswrapper[4692]: I0309 09:57:25.636934 4692 generic.go:334] "Generic (PLEG): container finished" podID="68efd13e-278f-4166-bf74-64e0cf2b5b9a" containerID="9e74e40ad1920b57ede3cf39989dae9ba4843dc6b52db3800f65d3b52cfe29b4" exitCode=0 Mar 09 09:57:25 crc kubenswrapper[4692]: I0309 09:57:25.636973 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" event={"ID":"68efd13e-278f-4166-bf74-64e0cf2b5b9a","Type":"ContainerDied","Data":"9e74e40ad1920b57ede3cf39989dae9ba4843dc6b52db3800f65d3b52cfe29b4"} Mar 09 09:57:26 crc kubenswrapper[4692]: I0309 09:57:26.906904 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:26 crc kubenswrapper[4692]: I0309 09:57:26.941389 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-69vn9"] Mar 09 09:57:26 crc kubenswrapper[4692]: I0309 09:57:26.948024 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-69vn9"] Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.007989 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-scripts\") pod \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.008071 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-ring-data-devices\") pod \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.008142 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-swiftconf\") pod \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.008214 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68efd13e-278f-4166-bf74-64e0cf2b5b9a-etc-swift\") pod \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.008232 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvk7r\" (UniqueName: \"kubernetes.io/projected/68efd13e-278f-4166-bf74-64e0cf2b5b9a-kube-api-access-gvk7r\") pod \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.008267 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-dispersionconf\") pod \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\" (UID: \"68efd13e-278f-4166-bf74-64e0cf2b5b9a\") " Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.009139 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "68efd13e-278f-4166-bf74-64e0cf2b5b9a" (UID: "68efd13e-278f-4166-bf74-64e0cf2b5b9a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.009397 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68efd13e-278f-4166-bf74-64e0cf2b5b9a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "68efd13e-278f-4166-bf74-64e0cf2b5b9a" (UID: "68efd13e-278f-4166-bf74-64e0cf2b5b9a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.015507 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68efd13e-278f-4166-bf74-64e0cf2b5b9a-kube-api-access-gvk7r" (OuterVolumeSpecName: "kube-api-access-gvk7r") pod "68efd13e-278f-4166-bf74-64e0cf2b5b9a" (UID: "68efd13e-278f-4166-bf74-64e0cf2b5b9a"). InnerVolumeSpecName "kube-api-access-gvk7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.027103 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-scripts" (OuterVolumeSpecName: "scripts") pod "68efd13e-278f-4166-bf74-64e0cf2b5b9a" (UID: "68efd13e-278f-4166-bf74-64e0cf2b5b9a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.029630 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "68efd13e-278f-4166-bf74-64e0cf2b5b9a" (UID: "68efd13e-278f-4166-bf74-64e0cf2b5b9a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.030616 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "68efd13e-278f-4166-bf74-64e0cf2b5b9a" (UID: "68efd13e-278f-4166-bf74-64e0cf2b5b9a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.109749 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.109807 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.109825 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/68efd13e-278f-4166-bf74-64e0cf2b5b9a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.109842 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/68efd13e-278f-4166-bf74-64e0cf2b5b9a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.109853 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/68efd13e-278f-4166-bf74-64e0cf2b5b9a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.109863 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvk7r\" (UniqueName: \"kubernetes.io/projected/68efd13e-278f-4166-bf74-64e0cf2b5b9a-kube-api-access-gvk7r\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.676529 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9316cbb361e3796954fb1e2f4c2d09085a99ad1e120a15ae48f14e27f4d7f086" Mar 09 09:57:27 crc kubenswrapper[4692]: I0309 09:57:27.676592 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-69vn9" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.092190 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68efd13e-278f-4166-bf74-64e0cf2b5b9a" path="/var/lib/kubelet/pods/68efd13e-278f-4166-bf74-64e0cf2b5b9a/volumes" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.106333 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2"] Mar 09 09:57:28 crc kubenswrapper[4692]: E0309 09:57:28.106641 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68efd13e-278f-4166-bf74-64e0cf2b5b9a" containerName="swift-ring-rebalance" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.106657 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="68efd13e-278f-4166-bf74-64e0cf2b5b9a" containerName="swift-ring-rebalance" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.106797 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="68efd13e-278f-4166-bf74-64e0cf2b5b9a" containerName="swift-ring-rebalance" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.107267 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.109735 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.109761 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.112828 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2"] Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.124375 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.124438 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-swiftconf\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.124554 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-etc-swift\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.124592 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-dispersionconf\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.124669 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-scripts\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.124772 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjbq4\" (UniqueName: \"kubernetes.io/projected/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-kube-api-access-kjbq4\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.226455 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjbq4\" (UniqueName: \"kubernetes.io/projected/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-kube-api-access-kjbq4\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.226504 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.226533 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-swiftconf\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.226578 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-etc-swift\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.226593 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-dispersionconf\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.226611 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-scripts\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.227327 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-scripts\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.227566 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-etc-swift\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.227670 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-ring-data-devices\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.234507 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-dispersionconf\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.239090 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-swiftconf\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.247401 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjbq4\" (UniqueName: \"kubernetes.io/projected/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-kube-api-access-kjbq4\") pod \"swift-ring-rebalance-debug-j4gz2\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.433852 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:28 crc kubenswrapper[4692]: I0309 09:57:28.869573 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2"] Mar 09 09:57:28 crc kubenswrapper[4692]: W0309 09:57:28.874787 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fe7c97f_0dde_477c_acea_b3a8e01dc5a6.slice/crio-6cc7b252dce92127202a9119f9ca2ec25380483fcf3e1001dd0b0bac03f061b8 WatchSource:0}: Error finding container 6cc7b252dce92127202a9119f9ca2ec25380483fcf3e1001dd0b0bac03f061b8: Status 404 returned error can't find the container with id 6cc7b252dce92127202a9119f9ca2ec25380483fcf3e1001dd0b0bac03f061b8 Mar 09 09:57:29 crc kubenswrapper[4692]: I0309 09:57:29.694789 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" event={"ID":"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6","Type":"ContainerStarted","Data":"c3b78fe422f72e5a6467bf1ec2e5ae0fce32725b991608efea641314fac1fa87"} Mar 09 09:57:29 crc kubenswrapper[4692]: I0309 09:57:29.695044 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" event={"ID":"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6","Type":"ContainerStarted","Data":"6cc7b252dce92127202a9119f9ca2ec25380483fcf3e1001dd0b0bac03f061b8"} Mar 09 09:57:30 crc kubenswrapper[4692]: I0309 09:57:30.075205 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:57:30 crc kubenswrapper[4692]: E0309 09:57:30.075434 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:57:30 crc kubenswrapper[4692]: I0309 09:57:30.706143 4692 generic.go:334] "Generic (PLEG): container finished" podID="8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" containerID="c3b78fe422f72e5a6467bf1ec2e5ae0fce32725b991608efea641314fac1fa87" exitCode=0 Mar 09 09:57:30 crc kubenswrapper[4692]: I0309 09:57:30.706242 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" event={"ID":"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6","Type":"ContainerDied","Data":"c3b78fe422f72e5a6467bf1ec2e5ae0fce32725b991608efea641314fac1fa87"} Mar 09 09:57:31 crc kubenswrapper[4692]: I0309 09:57:31.980007 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.009950 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2"] Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.016193 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2"] Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.082980 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-scripts\") pod \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.083034 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjbq4\" (UniqueName: \"kubernetes.io/projected/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-kube-api-access-kjbq4\") pod \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.083084 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-etc-swift\") pod \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.083152 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-ring-data-devices\") pod \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.083226 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-swiftconf\") pod \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.083250 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-dispersionconf\") pod \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\" (UID: \"8fe7c97f-0dde-477c-acea-b3a8e01dc5a6\") " Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.084350 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" (UID: "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.084550 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" (UID: "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.090364 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-kube-api-access-kjbq4" (OuterVolumeSpecName: "kube-api-access-kjbq4") pod "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" (UID: "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6"). InnerVolumeSpecName "kube-api-access-kjbq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.105827 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" (UID: "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.106629 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-scripts" (OuterVolumeSpecName: "scripts") pod "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" (UID: "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.112121 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" (UID: "8fe7c97f-0dde-477c-acea-b3a8e01dc5a6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.185904 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.186062 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.186096 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.186111 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.186128 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjbq4\" (UniqueName: \"kubernetes.io/projected/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-kube-api-access-kjbq4\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.186171 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.725312 4692 scope.go:117] "RemoveContainer" containerID="c3b78fe422f72e5a6467bf1ec2e5ae0fce32725b991608efea641314fac1fa87" Mar 09 09:57:32 crc kubenswrapper[4692]: I0309 09:57:32.725339 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j4gz2" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.171752 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq"] Mar 09 09:57:33 crc kubenswrapper[4692]: E0309 09:57:33.172263 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" containerName="swift-ring-rebalance" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.172281 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" containerName="swift-ring-rebalance" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.172468 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" containerName="swift-ring-rebalance" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.173061 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.179031 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq"] Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.222405 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.223773 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.301465 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h59vv\" (UniqueName: \"kubernetes.io/projected/83efb31e-4a90-4e1a-9d1c-62324130c1ca-kube-api-access-h59vv\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.301933 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-scripts\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.301965 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-ring-data-devices\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.301986 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-swiftconf\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.302053 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83efb31e-4a90-4e1a-9d1c-62324130c1ca-etc-swift\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.302113 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-dispersionconf\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.403626 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h59vv\" (UniqueName: \"kubernetes.io/projected/83efb31e-4a90-4e1a-9d1c-62324130c1ca-kube-api-access-h59vv\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.403702 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-ring-data-devices\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.403731 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-scripts\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.403756 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-swiftconf\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.403814 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83efb31e-4a90-4e1a-9d1c-62324130c1ca-etc-swift\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.403890 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-dispersionconf\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.404525 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83efb31e-4a90-4e1a-9d1c-62324130c1ca-etc-swift\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.404605 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-scripts\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.405238 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-ring-data-devices\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.411756 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-swiftconf\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.412663 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-dispersionconf\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.422083 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h59vv\" (UniqueName: \"kubernetes.io/projected/83efb31e-4a90-4e1a-9d1c-62324130c1ca-kube-api-access-h59vv\") pod \"swift-ring-rebalance-debug-wm4sq\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.545967 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:33 crc kubenswrapper[4692]: I0309 09:57:33.948330 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq"] Mar 09 09:57:33 crc kubenswrapper[4692]: W0309 09:57:33.950757 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83efb31e_4a90_4e1a_9d1c_62324130c1ca.slice/crio-8de9b5f76e29bf41edc93eb9813c982a5b3b9cbf9933266dc3d4312ae316afa3 WatchSource:0}: Error finding container 8de9b5f76e29bf41edc93eb9813c982a5b3b9cbf9933266dc3d4312ae316afa3: Status 404 returned error can't find the container with id 8de9b5f76e29bf41edc93eb9813c982a5b3b9cbf9933266dc3d4312ae316afa3 Mar 09 09:57:34 crc kubenswrapper[4692]: I0309 09:57:34.079811 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe7c97f-0dde-477c-acea-b3a8e01dc5a6" path="/var/lib/kubelet/pods/8fe7c97f-0dde-477c-acea-b3a8e01dc5a6/volumes" Mar 09 09:57:34 crc kubenswrapper[4692]: I0309 09:57:34.744225 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" event={"ID":"83efb31e-4a90-4e1a-9d1c-62324130c1ca","Type":"ContainerStarted","Data":"53b952ace5e9f05efb011c7e0e6bf396f45cc2ae4e531b8bd2bf38a1c7160c67"} Mar 09 09:57:34 crc kubenswrapper[4692]: I0309 09:57:34.744271 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" event={"ID":"83efb31e-4a90-4e1a-9d1c-62324130c1ca","Type":"ContainerStarted","Data":"8de9b5f76e29bf41edc93eb9813c982a5b3b9cbf9933266dc3d4312ae316afa3"} Mar 09 09:57:34 crc kubenswrapper[4692]: I0309 09:57:34.766517 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" podStartSLOduration=1.766501738 podStartE2EDuration="1.766501738s" podCreationTimestamp="2026-03-09 09:57:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:34.758930854 +0000 UTC m=+2255.583666445" watchObservedRunningTime="2026-03-09 09:57:34.766501738 +0000 UTC m=+2255.591237319" Mar 09 09:57:35 crc kubenswrapper[4692]: I0309 09:57:35.754908 4692 generic.go:334] "Generic (PLEG): container finished" podID="83efb31e-4a90-4e1a-9d1c-62324130c1ca" containerID="53b952ace5e9f05efb011c7e0e6bf396f45cc2ae4e531b8bd2bf38a1c7160c67" exitCode=0 Mar 09 09:57:35 crc kubenswrapper[4692]: I0309 09:57:35.754964 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" event={"ID":"83efb31e-4a90-4e1a-9d1c-62324130c1ca","Type":"ContainerDied","Data":"53b952ace5e9f05efb011c7e0e6bf396f45cc2ae4e531b8bd2bf38a1c7160c67"} Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.046408 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.091573 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq"] Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.099060 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq"] Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.159041 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-ring-data-devices\") pod \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.159403 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-scripts\") pod \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.159606 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83efb31e-4a90-4e1a-9d1c-62324130c1ca-etc-swift\") pod \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.159700 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-dispersionconf\") pod \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.159794 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-swiftconf\") pod \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.159923 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h59vv\" (UniqueName: \"kubernetes.io/projected/83efb31e-4a90-4e1a-9d1c-62324130c1ca-kube-api-access-h59vv\") pod \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\" (UID: \"83efb31e-4a90-4e1a-9d1c-62324130c1ca\") " Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.159984 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "83efb31e-4a90-4e1a-9d1c-62324130c1ca" (UID: "83efb31e-4a90-4e1a-9d1c-62324130c1ca"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.160303 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.160441 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83efb31e-4a90-4e1a-9d1c-62324130c1ca-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "83efb31e-4a90-4e1a-9d1c-62324130c1ca" (UID: "83efb31e-4a90-4e1a-9d1c-62324130c1ca"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.166457 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83efb31e-4a90-4e1a-9d1c-62324130c1ca-kube-api-access-h59vv" (OuterVolumeSpecName: "kube-api-access-h59vv") pod "83efb31e-4a90-4e1a-9d1c-62324130c1ca" (UID: "83efb31e-4a90-4e1a-9d1c-62324130c1ca"). InnerVolumeSpecName "kube-api-access-h59vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.180739 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-scripts" (OuterVolumeSpecName: "scripts") pod "83efb31e-4a90-4e1a-9d1c-62324130c1ca" (UID: "83efb31e-4a90-4e1a-9d1c-62324130c1ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.182120 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "83efb31e-4a90-4e1a-9d1c-62324130c1ca" (UID: "83efb31e-4a90-4e1a-9d1c-62324130c1ca"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.182735 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "83efb31e-4a90-4e1a-9d1c-62324130c1ca" (UID: "83efb31e-4a90-4e1a-9d1c-62324130c1ca"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.261560 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.261606 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h59vv\" (UniqueName: \"kubernetes.io/projected/83efb31e-4a90-4e1a-9d1c-62324130c1ca-kube-api-access-h59vv\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.261620 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83efb31e-4a90-4e1a-9d1c-62324130c1ca-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.261633 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83efb31e-4a90-4e1a-9d1c-62324130c1ca-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.261645 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83efb31e-4a90-4e1a-9d1c-62324130c1ca-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.774246 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8de9b5f76e29bf41edc93eb9813c982a5b3b9cbf9933266dc3d4312ae316afa3" Mar 09 09:57:37 crc kubenswrapper[4692]: I0309 09:57:37.774318 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wm4sq" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.080815 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83efb31e-4a90-4e1a-9d1c-62324130c1ca" path="/var/lib/kubelet/pods/83efb31e-4a90-4e1a-9d1c-62324130c1ca/volumes" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.259804 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zqclx"] Mar 09 09:57:38 crc kubenswrapper[4692]: E0309 09:57:38.260449 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83efb31e-4a90-4e1a-9d1c-62324130c1ca" containerName="swift-ring-rebalance" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.260463 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="83efb31e-4a90-4e1a-9d1c-62324130c1ca" containerName="swift-ring-rebalance" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.260659 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="83efb31e-4a90-4e1a-9d1c-62324130c1ca" containerName="swift-ring-rebalance" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.261373 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.264496 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.265253 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.267839 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zqclx"] Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.375546 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-ring-data-devices\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.375589 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7zmj\" (UniqueName: \"kubernetes.io/projected/40448d25-63f5-4223-a99e-9932fbb9d03a-kube-api-access-n7zmj\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.375638 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40448d25-63f5-4223-a99e-9932fbb9d03a-etc-swift\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.376146 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-scripts\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.376236 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-dispersionconf\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.376332 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-swiftconf\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.477699 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-swiftconf\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.477759 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-ring-data-devices\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.477811 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7zmj\" (UniqueName: \"kubernetes.io/projected/40448d25-63f5-4223-a99e-9932fbb9d03a-kube-api-access-n7zmj\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.477839 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40448d25-63f5-4223-a99e-9932fbb9d03a-etc-swift\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.477883 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-scripts\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.477909 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-dispersionconf\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.478513 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40448d25-63f5-4223-a99e-9932fbb9d03a-etc-swift\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.478850 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-scripts\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.479002 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-ring-data-devices\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.482804 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-dispersionconf\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.483743 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-swiftconf\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.494137 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7zmj\" (UniqueName: \"kubernetes.io/projected/40448d25-63f5-4223-a99e-9932fbb9d03a-kube-api-access-n7zmj\") pod \"swift-ring-rebalance-debug-zqclx\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:38 crc kubenswrapper[4692]: I0309 09:57:38.581798 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:39 crc kubenswrapper[4692]: I0309 09:57:39.055001 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zqclx"] Mar 09 09:57:39 crc kubenswrapper[4692]: I0309 09:57:39.805647 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" event={"ID":"40448d25-63f5-4223-a99e-9932fbb9d03a","Type":"ContainerStarted","Data":"934b15b36056a6e0d51982da677933fd90ed6ecdec82a15b2d71712f3a7a696f"} Mar 09 09:57:39 crc kubenswrapper[4692]: I0309 09:57:39.805696 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" event={"ID":"40448d25-63f5-4223-a99e-9932fbb9d03a","Type":"ContainerStarted","Data":"6c41c23771a491bb958a719560a15314b59a0bbf45434edd2bd49d016ab1ca56"} Mar 09 09:57:40 crc kubenswrapper[4692]: I0309 09:57:40.815057 4692 generic.go:334] "Generic (PLEG): container finished" podID="40448d25-63f5-4223-a99e-9932fbb9d03a" containerID="934b15b36056a6e0d51982da677933fd90ed6ecdec82a15b2d71712f3a7a696f" exitCode=0 Mar 09 09:57:40 crc kubenswrapper[4692]: I0309 09:57:40.815107 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" event={"ID":"40448d25-63f5-4223-a99e-9932fbb9d03a","Type":"ContainerDied","Data":"934b15b36056a6e0d51982da677933fd90ed6ecdec82a15b2d71712f3a7a696f"} Mar 09 09:57:41 crc kubenswrapper[4692]: I0309 09:57:41.070879 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:57:41 crc kubenswrapper[4692]: E0309 09:57:41.071177 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.131342 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.165436 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zqclx"] Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.170433 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-zqclx"] Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.171782 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-swiftconf\") pod \"40448d25-63f5-4223-a99e-9932fbb9d03a\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.171870 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7zmj\" (UniqueName: \"kubernetes.io/projected/40448d25-63f5-4223-a99e-9932fbb9d03a-kube-api-access-n7zmj\") pod \"40448d25-63f5-4223-a99e-9932fbb9d03a\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.171894 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-scripts\") pod \"40448d25-63f5-4223-a99e-9932fbb9d03a\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.171919 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40448d25-63f5-4223-a99e-9932fbb9d03a-etc-swift\") pod \"40448d25-63f5-4223-a99e-9932fbb9d03a\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.171945 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-dispersionconf\") pod \"40448d25-63f5-4223-a99e-9932fbb9d03a\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.172022 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-ring-data-devices\") pod \"40448d25-63f5-4223-a99e-9932fbb9d03a\" (UID: \"40448d25-63f5-4223-a99e-9932fbb9d03a\") " Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.173428 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "40448d25-63f5-4223-a99e-9932fbb9d03a" (UID: "40448d25-63f5-4223-a99e-9932fbb9d03a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.174038 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40448d25-63f5-4223-a99e-9932fbb9d03a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "40448d25-63f5-4223-a99e-9932fbb9d03a" (UID: "40448d25-63f5-4223-a99e-9932fbb9d03a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.176999 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40448d25-63f5-4223-a99e-9932fbb9d03a-kube-api-access-n7zmj" (OuterVolumeSpecName: "kube-api-access-n7zmj") pod "40448d25-63f5-4223-a99e-9932fbb9d03a" (UID: "40448d25-63f5-4223-a99e-9932fbb9d03a"). InnerVolumeSpecName "kube-api-access-n7zmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.190591 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-scripts" (OuterVolumeSpecName: "scripts") pod "40448d25-63f5-4223-a99e-9932fbb9d03a" (UID: "40448d25-63f5-4223-a99e-9932fbb9d03a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.194055 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "40448d25-63f5-4223-a99e-9932fbb9d03a" (UID: "40448d25-63f5-4223-a99e-9932fbb9d03a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.195115 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "40448d25-63f5-4223-a99e-9932fbb9d03a" (UID: "40448d25-63f5-4223-a99e-9932fbb9d03a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.274631 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.274678 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.274689 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40448d25-63f5-4223-a99e-9932fbb9d03a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.274698 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7zmj\" (UniqueName: \"kubernetes.io/projected/40448d25-63f5-4223-a99e-9932fbb9d03a-kube-api-access-n7zmj\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.274710 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40448d25-63f5-4223-a99e-9932fbb9d03a-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.274718 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40448d25-63f5-4223-a99e-9932fbb9d03a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.833196 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c41c23771a491bb958a719560a15314b59a0bbf45434edd2bd49d016ab1ca56" Mar 09 09:57:42 crc kubenswrapper[4692]: I0309 09:57:42.833515 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-zqclx" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.307782 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kf86"] Mar 09 09:57:43 crc kubenswrapper[4692]: E0309 09:57:43.309628 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40448d25-63f5-4223-a99e-9932fbb9d03a" containerName="swift-ring-rebalance" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.309737 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="40448d25-63f5-4223-a99e-9932fbb9d03a" containerName="swift-ring-rebalance" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.310103 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="40448d25-63f5-4223-a99e-9932fbb9d03a" containerName="swift-ring-rebalance" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.310855 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.314872 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.315198 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.322111 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kf86"] Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.392208 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f6028d3-5da3-415f-93ae-9db1bc11112e-etc-swift\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.392614 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-scripts\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.392725 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-ring-data-devices\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.392870 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjbs7\" (UniqueName: \"kubernetes.io/projected/3f6028d3-5da3-415f-93ae-9db1bc11112e-kube-api-access-fjbs7\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.392973 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-dispersionconf\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.393064 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-swiftconf\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.494110 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjbs7\" (UniqueName: \"kubernetes.io/projected/3f6028d3-5da3-415f-93ae-9db1bc11112e-kube-api-access-fjbs7\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.494228 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-dispersionconf\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.494264 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-swiftconf\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.494293 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f6028d3-5da3-415f-93ae-9db1bc11112e-etc-swift\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.494310 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-scripts\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.494354 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-ring-data-devices\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.494922 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f6028d3-5da3-415f-93ae-9db1bc11112e-etc-swift\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.495123 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-ring-data-devices\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.495340 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-scripts\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.499997 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-dispersionconf\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.501303 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-swiftconf\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.510620 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjbs7\" (UniqueName: \"kubernetes.io/projected/3f6028d3-5da3-415f-93ae-9db1bc11112e-kube-api-access-fjbs7\") pod \"swift-ring-rebalance-debug-4kf86\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:43 crc kubenswrapper[4692]: I0309 09:57:43.637335 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:44 crc kubenswrapper[4692]: I0309 09:57:44.086860 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40448d25-63f5-4223-a99e-9932fbb9d03a" path="/var/lib/kubelet/pods/40448d25-63f5-4223-a99e-9932fbb9d03a/volumes" Mar 09 09:57:44 crc kubenswrapper[4692]: I0309 09:57:44.092773 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kf86"] Mar 09 09:57:44 crc kubenswrapper[4692]: I0309 09:57:44.851674 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" event={"ID":"3f6028d3-5da3-415f-93ae-9db1bc11112e","Type":"ContainerStarted","Data":"fff1aa1f2b5e7d05320624b5bef883ca44ed6de52835fb6652c1ea1c0dcfd75c"} Mar 09 09:57:44 crc kubenswrapper[4692]: I0309 09:57:44.852148 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" event={"ID":"3f6028d3-5da3-415f-93ae-9db1bc11112e","Type":"ContainerStarted","Data":"6b585169c59ab1589fd04ef75cb448dcc2ed6ab9896850ce8c2c91c07bd7f620"} Mar 09 09:57:44 crc kubenswrapper[4692]: I0309 09:57:44.880250 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" podStartSLOduration=1.880151328 podStartE2EDuration="1.880151328s" podCreationTimestamp="2026-03-09 09:57:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:44.875692742 +0000 UTC m=+2265.700428373" watchObservedRunningTime="2026-03-09 09:57:44.880151328 +0000 UTC m=+2265.704886939" Mar 09 09:57:45 crc kubenswrapper[4692]: I0309 09:57:45.861571 4692 generic.go:334] "Generic (PLEG): container finished" podID="3f6028d3-5da3-415f-93ae-9db1bc11112e" containerID="fff1aa1f2b5e7d05320624b5bef883ca44ed6de52835fb6652c1ea1c0dcfd75c" exitCode=0 Mar 09 09:57:45 crc kubenswrapper[4692]: I0309 09:57:45.861625 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" event={"ID":"3f6028d3-5da3-415f-93ae-9db1bc11112e","Type":"ContainerDied","Data":"fff1aa1f2b5e7d05320624b5bef883ca44ed6de52835fb6652c1ea1c0dcfd75c"} Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.127137 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.167739 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kf86"] Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.174816 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kf86"] Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.258998 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-dispersionconf\") pod \"3f6028d3-5da3-415f-93ae-9db1bc11112e\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.259047 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-swiftconf\") pod \"3f6028d3-5da3-415f-93ae-9db1bc11112e\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.259125 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f6028d3-5da3-415f-93ae-9db1bc11112e-etc-swift\") pod \"3f6028d3-5da3-415f-93ae-9db1bc11112e\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.259974 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6028d3-5da3-415f-93ae-9db1bc11112e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3f6028d3-5da3-415f-93ae-9db1bc11112e" (UID: "3f6028d3-5da3-415f-93ae-9db1bc11112e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.260060 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-ring-data-devices\") pod \"3f6028d3-5da3-415f-93ae-9db1bc11112e\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.260660 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3f6028d3-5da3-415f-93ae-9db1bc11112e" (UID: "3f6028d3-5da3-415f-93ae-9db1bc11112e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.260823 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjbs7\" (UniqueName: \"kubernetes.io/projected/3f6028d3-5da3-415f-93ae-9db1bc11112e-kube-api-access-fjbs7\") pod \"3f6028d3-5da3-415f-93ae-9db1bc11112e\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.260919 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-scripts\") pod \"3f6028d3-5da3-415f-93ae-9db1bc11112e\" (UID: \"3f6028d3-5da3-415f-93ae-9db1bc11112e\") " Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.261403 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f6028d3-5da3-415f-93ae-9db1bc11112e-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.261422 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.267032 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f6028d3-5da3-415f-93ae-9db1bc11112e-kube-api-access-fjbs7" (OuterVolumeSpecName: "kube-api-access-fjbs7") pod "3f6028d3-5da3-415f-93ae-9db1bc11112e" (UID: "3f6028d3-5da3-415f-93ae-9db1bc11112e"). InnerVolumeSpecName "kube-api-access-fjbs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.283927 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-scripts" (OuterVolumeSpecName: "scripts") pod "3f6028d3-5da3-415f-93ae-9db1bc11112e" (UID: "3f6028d3-5da3-415f-93ae-9db1bc11112e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.288674 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3f6028d3-5da3-415f-93ae-9db1bc11112e" (UID: "3f6028d3-5da3-415f-93ae-9db1bc11112e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.289474 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3f6028d3-5da3-415f-93ae-9db1bc11112e" (UID: "3f6028d3-5da3-415f-93ae-9db1bc11112e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.363384 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.363435 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f6028d3-5da3-415f-93ae-9db1bc11112e-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.363449 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjbs7\" (UniqueName: \"kubernetes.io/projected/3f6028d3-5da3-415f-93ae-9db1bc11112e-kube-api-access-fjbs7\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.363466 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6028d3-5da3-415f-93ae-9db1bc11112e-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.882206 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b585169c59ab1589fd04ef75cb448dcc2ed6ab9896850ce8c2c91c07bd7f620" Mar 09 09:57:47 crc kubenswrapper[4692]: I0309 09:57:47.882266 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kf86" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.081439 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f6028d3-5da3-415f-93ae-9db1bc11112e" path="/var/lib/kubelet/pods/3f6028d3-5da3-415f-93ae-9db1bc11112e/volumes" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.314522 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qw48m"] Mar 09 09:57:48 crc kubenswrapper[4692]: E0309 09:57:48.315221 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6028d3-5da3-415f-93ae-9db1bc11112e" containerName="swift-ring-rebalance" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.315234 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6028d3-5da3-415f-93ae-9db1bc11112e" containerName="swift-ring-rebalance" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.315385 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6028d3-5da3-415f-93ae-9db1bc11112e" containerName="swift-ring-rebalance" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.315852 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.319530 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.319781 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.341791 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qw48m"] Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.377811 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-ring-data-devices\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.378005 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t758s\" (UniqueName: \"kubernetes.io/projected/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-kube-api-access-t758s\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.378134 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-scripts\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.378236 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-dispersionconf\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.378292 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-swiftconf\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.378352 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-etc-swift\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.479739 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-ring-data-devices\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.479847 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t758s\" (UniqueName: \"kubernetes.io/projected/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-kube-api-access-t758s\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.479900 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-scripts\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.479923 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-dispersionconf\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.479946 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-swiftconf\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.479974 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-etc-swift\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.480618 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-etc-swift\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.480694 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-ring-data-devices\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.480793 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-scripts\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.484638 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-swiftconf\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.497669 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-dispersionconf\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.500296 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t758s\" (UniqueName: \"kubernetes.io/projected/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-kube-api-access-t758s\") pod \"swift-ring-rebalance-debug-qw48m\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:48 crc kubenswrapper[4692]: I0309 09:57:48.658063 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:49 crc kubenswrapper[4692]: I0309 09:57:49.129400 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qw48m"] Mar 09 09:57:49 crc kubenswrapper[4692]: I0309 09:57:49.929091 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" event={"ID":"8f9eaf8a-739b-4f4f-9354-c9b968faeb46","Type":"ContainerStarted","Data":"7dfaaa7954a17d8a987b64ba7c2dcec0ad6c6676b10c7fbefb94f71b6a62d9a8"} Mar 09 09:57:49 crc kubenswrapper[4692]: I0309 09:57:49.929544 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" event={"ID":"8f9eaf8a-739b-4f4f-9354-c9b968faeb46","Type":"ContainerStarted","Data":"dceaba9f776045a95515a3f6d20fc8465e35468475127d6b1558841fb229774d"} Mar 09 09:57:49 crc kubenswrapper[4692]: I0309 09:57:49.954293 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" podStartSLOduration=1.954150064 podStartE2EDuration="1.954150064s" podCreationTimestamp="2026-03-09 09:57:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:49.948150605 +0000 UTC m=+2270.772886236" watchObservedRunningTime="2026-03-09 09:57:49.954150064 +0000 UTC m=+2270.778885655" Mar 09 09:57:50 crc kubenswrapper[4692]: I0309 09:57:50.938667 4692 generic.go:334] "Generic (PLEG): container finished" podID="8f9eaf8a-739b-4f4f-9354-c9b968faeb46" containerID="7dfaaa7954a17d8a987b64ba7c2dcec0ad6c6676b10c7fbefb94f71b6a62d9a8" exitCode=0 Mar 09 09:57:50 crc kubenswrapper[4692]: I0309 09:57:50.939756 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" event={"ID":"8f9eaf8a-739b-4f4f-9354-c9b968faeb46","Type":"ContainerDied","Data":"7dfaaa7954a17d8a987b64ba7c2dcec0ad6c6676b10c7fbefb94f71b6a62d9a8"} Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.240013 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.288775 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qw48m"] Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.296370 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qw48m"] Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.359388 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-etc-swift\") pod \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.359483 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-dispersionconf\") pod \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.359534 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-ring-data-devices\") pod \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.359576 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-scripts\") pod \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.359603 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-swiftconf\") pod \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.359693 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t758s\" (UniqueName: \"kubernetes.io/projected/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-kube-api-access-t758s\") pod \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\" (UID: \"8f9eaf8a-739b-4f4f-9354-c9b968faeb46\") " Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.360073 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8f9eaf8a-739b-4f4f-9354-c9b968faeb46" (UID: "8f9eaf8a-739b-4f4f-9354-c9b968faeb46"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.360458 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8f9eaf8a-739b-4f4f-9354-c9b968faeb46" (UID: "8f9eaf8a-739b-4f4f-9354-c9b968faeb46"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.369500 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-kube-api-access-t758s" (OuterVolumeSpecName: "kube-api-access-t758s") pod "8f9eaf8a-739b-4f4f-9354-c9b968faeb46" (UID: "8f9eaf8a-739b-4f4f-9354-c9b968faeb46"). InnerVolumeSpecName "kube-api-access-t758s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.380875 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-scripts" (OuterVolumeSpecName: "scripts") pod "8f9eaf8a-739b-4f4f-9354-c9b968faeb46" (UID: "8f9eaf8a-739b-4f4f-9354-c9b968faeb46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.383005 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8f9eaf8a-739b-4f4f-9354-c9b968faeb46" (UID: "8f9eaf8a-739b-4f4f-9354-c9b968faeb46"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.387465 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8f9eaf8a-739b-4f4f-9354-c9b968faeb46" (UID: "8f9eaf8a-739b-4f4f-9354-c9b968faeb46"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.461391 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.461431 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.461445 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.461459 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.461468 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.461477 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t758s\" (UniqueName: \"kubernetes.io/projected/8f9eaf8a-739b-4f4f-9354-c9b968faeb46-kube-api-access-t758s\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.966540 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dceaba9f776045a95515a3f6d20fc8465e35468475127d6b1558841fb229774d" Mar 09 09:57:52 crc kubenswrapper[4692]: I0309 09:57:52.966666 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qw48m" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.461774 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-chq2j"] Mar 09 09:57:53 crc kubenswrapper[4692]: E0309 09:57:53.462527 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f9eaf8a-739b-4f4f-9354-c9b968faeb46" containerName="swift-ring-rebalance" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.462548 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f9eaf8a-739b-4f4f-9354-c9b968faeb46" containerName="swift-ring-rebalance" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.462701 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f9eaf8a-739b-4f4f-9354-c9b968faeb46" containerName="swift-ring-rebalance" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.463353 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.466569 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.466868 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.476237 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-chq2j"] Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.579773 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-swiftconf\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.579840 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcqhv\" (UniqueName: \"kubernetes.io/projected/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-kube-api-access-wcqhv\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.579870 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-ring-data-devices\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.579917 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-etc-swift\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.580052 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-dispersionconf\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.580375 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-scripts\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.681893 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-swiftconf\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.681964 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcqhv\" (UniqueName: \"kubernetes.io/projected/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-kube-api-access-wcqhv\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.681994 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-ring-data-devices\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.682040 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-etc-swift\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.682084 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-dispersionconf\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.682133 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-scripts\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.683064 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-scripts\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.685482 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-etc-swift\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.685521 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-ring-data-devices\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.690427 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-dispersionconf\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.691208 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-swiftconf\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.703638 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcqhv\" (UniqueName: \"kubernetes.io/projected/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-kube-api-access-wcqhv\") pod \"swift-ring-rebalance-debug-chq2j\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:53 crc kubenswrapper[4692]: I0309 09:57:53.788890 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:54 crc kubenswrapper[4692]: I0309 09:57:54.079724 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9eaf8a-739b-4f4f-9354-c9b968faeb46" path="/var/lib/kubelet/pods/8f9eaf8a-739b-4f4f-9354-c9b968faeb46/volumes" Mar 09 09:57:54 crc kubenswrapper[4692]: I0309 09:57:54.227608 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-chq2j"] Mar 09 09:57:54 crc kubenswrapper[4692]: W0309 09:57:54.235587 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26d4a35a_c415_4923_91c9_d9a1a1a2e48c.slice/crio-78f2016cc8216c13fec5db196fa143aebb25873badb42e604f4a36b034836190 WatchSource:0}: Error finding container 78f2016cc8216c13fec5db196fa143aebb25873badb42e604f4a36b034836190: Status 404 returned error can't find the container with id 78f2016cc8216c13fec5db196fa143aebb25873badb42e604f4a36b034836190 Mar 09 09:57:54 crc kubenswrapper[4692]: I0309 09:57:54.992533 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" event={"ID":"26d4a35a-c415-4923-91c9-d9a1a1a2e48c","Type":"ContainerStarted","Data":"44a1e7077d33ca5a2fb5c5b946c9afeb7a6332eeb50fc07e4b77cda6a08bfa99"} Mar 09 09:57:54 crc kubenswrapper[4692]: I0309 09:57:54.992965 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" event={"ID":"26d4a35a-c415-4923-91c9-d9a1a1a2e48c","Type":"ContainerStarted","Data":"78f2016cc8216c13fec5db196fa143aebb25873badb42e604f4a36b034836190"} Mar 09 09:57:55 crc kubenswrapper[4692]: I0309 09:57:55.017631 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" podStartSLOduration=2.017612862 podStartE2EDuration="2.017612862s" podCreationTimestamp="2026-03-09 09:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:57:55.011976533 +0000 UTC m=+2275.836712114" watchObservedRunningTime="2026-03-09 09:57:55.017612862 +0000 UTC m=+2275.842348443" Mar 09 09:57:56 crc kubenswrapper[4692]: I0309 09:57:56.002682 4692 generic.go:334] "Generic (PLEG): container finished" podID="26d4a35a-c415-4923-91c9-d9a1a1a2e48c" containerID="44a1e7077d33ca5a2fb5c5b946c9afeb7a6332eeb50fc07e4b77cda6a08bfa99" exitCode=0 Mar 09 09:57:56 crc kubenswrapper[4692]: I0309 09:57:56.002729 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" event={"ID":"26d4a35a-c415-4923-91c9-d9a1a1a2e48c","Type":"ContainerDied","Data":"44a1e7077d33ca5a2fb5c5b946c9afeb7a6332eeb50fc07e4b77cda6a08bfa99"} Mar 09 09:57:56 crc kubenswrapper[4692]: I0309 09:57:56.072019 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:57:56 crc kubenswrapper[4692]: E0309 09:57:56.072402 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.328521 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.373021 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-chq2j"] Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.381615 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-chq2j"] Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.450340 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-etc-swift\") pod \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.451254 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "26d4a35a-c415-4923-91c9-d9a1a1a2e48c" (UID: "26d4a35a-c415-4923-91c9-d9a1a1a2e48c"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.451425 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcqhv\" (UniqueName: \"kubernetes.io/projected/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-kube-api-access-wcqhv\") pod \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.451673 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-swiftconf\") pod \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.451772 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-ring-data-devices\") pod \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.451846 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-scripts\") pod \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.451894 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-dispersionconf\") pod \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\" (UID: \"26d4a35a-c415-4923-91c9-d9a1a1a2e48c\") " Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.452718 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.453049 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "26d4a35a-c415-4923-91c9-d9a1a1a2e48c" (UID: "26d4a35a-c415-4923-91c9-d9a1a1a2e48c"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.471549 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-kube-api-access-wcqhv" (OuterVolumeSpecName: "kube-api-access-wcqhv") pod "26d4a35a-c415-4923-91c9-d9a1a1a2e48c" (UID: "26d4a35a-c415-4923-91c9-d9a1a1a2e48c"). InnerVolumeSpecName "kube-api-access-wcqhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.480024 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-scripts" (OuterVolumeSpecName: "scripts") pod "26d4a35a-c415-4923-91c9-d9a1a1a2e48c" (UID: "26d4a35a-c415-4923-91c9-d9a1a1a2e48c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.480994 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "26d4a35a-c415-4923-91c9-d9a1a1a2e48c" (UID: "26d4a35a-c415-4923-91c9-d9a1a1a2e48c"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.483205 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "26d4a35a-c415-4923-91c9-d9a1a1a2e48c" (UID: "26d4a35a-c415-4923-91c9-d9a1a1a2e48c"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.554288 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcqhv\" (UniqueName: \"kubernetes.io/projected/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-kube-api-access-wcqhv\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.554341 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.554357 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.554369 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:57 crc kubenswrapper[4692]: I0309 09:57:57.554381 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26d4a35a-c415-4923-91c9-d9a1a1a2e48c-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.022840 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f2016cc8216c13fec5db196fa143aebb25873badb42e604f4a36b034836190" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.022878 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-chq2j" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.081105 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26d4a35a-c415-4923-91c9-d9a1a1a2e48c" path="/var/lib/kubelet/pods/26d4a35a-c415-4923-91c9-d9a1a1a2e48c/volumes" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.538989 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bgnql"] Mar 09 09:57:58 crc kubenswrapper[4692]: E0309 09:57:58.539477 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d4a35a-c415-4923-91c9-d9a1a1a2e48c" containerName="swift-ring-rebalance" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.539496 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d4a35a-c415-4923-91c9-d9a1a1a2e48c" containerName="swift-ring-rebalance" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.539725 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="26d4a35a-c415-4923-91c9-d9a1a1a2e48c" containerName="swift-ring-rebalance" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.540635 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.546680 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.546950 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.559836 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bgnql"] Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.674106 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-scripts\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.674283 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-swiftconf\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.674324 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v645m\" (UniqueName: \"kubernetes.io/projected/e3137730-712f-43ab-8c7c-bfd9640f3049-kube-api-access-v645m\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.674351 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-ring-data-devices\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.674766 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3137730-712f-43ab-8c7c-bfd9640f3049-etc-swift\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.674908 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-dispersionconf\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.775997 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3137730-712f-43ab-8c7c-bfd9640f3049-etc-swift\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.776055 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-dispersionconf\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.776096 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-scripts\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.776114 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-swiftconf\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.776157 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v645m\" (UniqueName: \"kubernetes.io/projected/e3137730-712f-43ab-8c7c-bfd9640f3049-kube-api-access-v645m\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.776312 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-ring-data-devices\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.776714 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3137730-712f-43ab-8c7c-bfd9640f3049-etc-swift\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.777218 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-scripts\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.777517 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-ring-data-devices\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.784837 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-swiftconf\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.794181 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-dispersionconf\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.797892 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v645m\" (UniqueName: \"kubernetes.io/projected/e3137730-712f-43ab-8c7c-bfd9640f3049-kube-api-access-v645m\") pod \"swift-ring-rebalance-debug-bgnql\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:58 crc kubenswrapper[4692]: I0309 09:57:58.871991 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:57:59 crc kubenswrapper[4692]: I0309 09:57:59.339418 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bgnql"] Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.049224 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" event={"ID":"e3137730-712f-43ab-8c7c-bfd9640f3049","Type":"ContainerStarted","Data":"8033743fc6419ba0a01e3a0e5cac0d473add21dc8f4bcfc2aac21723d388ff61"} Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.049307 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" event={"ID":"e3137730-712f-43ab-8c7c-bfd9640f3049","Type":"ContainerStarted","Data":"f5c6b0bfa2b281990271397a1aa2cc87a1122da41154188a543e531f438b8dd8"} Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.081338 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" podStartSLOduration=2.081312587 podStartE2EDuration="2.081312587s" podCreationTimestamp="2026-03-09 09:57:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:00.07612808 +0000 UTC m=+2280.900863661" watchObservedRunningTime="2026-03-09 09:58:00.081312587 +0000 UTC m=+2280.906048168" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.149518 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550838-9zpjw"] Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.150985 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.154575 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.154595 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.161524 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.169927 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550838-9zpjw"] Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.217736 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v96p9\" (UniqueName: \"kubernetes.io/projected/881ca87a-0f16-428d-b4e7-f9832b7b6f3e-kube-api-access-v96p9\") pod \"auto-csr-approver-29550838-9zpjw\" (UID: \"881ca87a-0f16-428d-b4e7-f9832b7b6f3e\") " pod="openshift-infra/auto-csr-approver-29550838-9zpjw" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.319555 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v96p9\" (UniqueName: \"kubernetes.io/projected/881ca87a-0f16-428d-b4e7-f9832b7b6f3e-kube-api-access-v96p9\") pod \"auto-csr-approver-29550838-9zpjw\" (UID: \"881ca87a-0f16-428d-b4e7-f9832b7b6f3e\") " pod="openshift-infra/auto-csr-approver-29550838-9zpjw" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.338710 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v96p9\" (UniqueName: \"kubernetes.io/projected/881ca87a-0f16-428d-b4e7-f9832b7b6f3e-kube-api-access-v96p9\") pod \"auto-csr-approver-29550838-9zpjw\" (UID: \"881ca87a-0f16-428d-b4e7-f9832b7b6f3e\") " pod="openshift-infra/auto-csr-approver-29550838-9zpjw" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.502838 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" Mar 09 09:58:00 crc kubenswrapper[4692]: I0309 09:58:00.766625 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550838-9zpjw"] Mar 09 09:58:01 crc kubenswrapper[4692]: I0309 09:58:01.060815 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" event={"ID":"881ca87a-0f16-428d-b4e7-f9832b7b6f3e","Type":"ContainerStarted","Data":"d8c1a12d12371e72852817490872e4f0f9b0ab7589c2cc13436161e6fdd7443d"} Mar 09 09:58:02 crc kubenswrapper[4692]: I0309 09:58:02.077105 4692 generic.go:334] "Generic (PLEG): container finished" podID="e3137730-712f-43ab-8c7c-bfd9640f3049" containerID="8033743fc6419ba0a01e3a0e5cac0d473add21dc8f4bcfc2aac21723d388ff61" exitCode=0 Mar 09 09:58:02 crc kubenswrapper[4692]: I0309 09:58:02.083267 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" event={"ID":"e3137730-712f-43ab-8c7c-bfd9640f3049","Type":"ContainerDied","Data":"8033743fc6419ba0a01e3a0e5cac0d473add21dc8f4bcfc2aac21723d388ff61"} Mar 09 09:58:02 crc kubenswrapper[4692]: I0309 09:58:02.083327 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" event={"ID":"881ca87a-0f16-428d-b4e7-f9832b7b6f3e","Type":"ContainerStarted","Data":"a4b7b2479f08482c464c10d79b90425486c8c45a8b1d8fb433f36e0d310223c6"} Mar 09 09:58:03 crc kubenswrapper[4692]: I0309 09:58:03.098181 4692 generic.go:334] "Generic (PLEG): container finished" podID="881ca87a-0f16-428d-b4e7-f9832b7b6f3e" containerID="a4b7b2479f08482c464c10d79b90425486c8c45a8b1d8fb433f36e0d310223c6" exitCode=0 Mar 09 09:58:03 crc kubenswrapper[4692]: I0309 09:58:03.098443 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" event={"ID":"881ca87a-0f16-428d-b4e7-f9832b7b6f3e","Type":"ContainerDied","Data":"a4b7b2479f08482c464c10d79b90425486c8c45a8b1d8fb433f36e0d310223c6"} Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.508306 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.521379 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.581074 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3137730-712f-43ab-8c7c-bfd9640f3049-etc-swift\") pod \"e3137730-712f-43ab-8c7c-bfd9640f3049\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.581193 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-swiftconf\") pod \"e3137730-712f-43ab-8c7c-bfd9640f3049\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.581487 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-ring-data-devices\") pod \"e3137730-712f-43ab-8c7c-bfd9640f3049\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.581544 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v96p9\" (UniqueName: \"kubernetes.io/projected/881ca87a-0f16-428d-b4e7-f9832b7b6f3e-kube-api-access-v96p9\") pod \"881ca87a-0f16-428d-b4e7-f9832b7b6f3e\" (UID: \"881ca87a-0f16-428d-b4e7-f9832b7b6f3e\") " Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.581581 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-dispersionconf\") pod \"e3137730-712f-43ab-8c7c-bfd9640f3049\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.581623 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-scripts\") pod \"e3137730-712f-43ab-8c7c-bfd9640f3049\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.582125 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3137730-712f-43ab-8c7c-bfd9640f3049-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e3137730-712f-43ab-8c7c-bfd9640f3049" (UID: "e3137730-712f-43ab-8c7c-bfd9640f3049"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.582529 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v645m\" (UniqueName: \"kubernetes.io/projected/e3137730-712f-43ab-8c7c-bfd9640f3049-kube-api-access-v645m\") pod \"e3137730-712f-43ab-8c7c-bfd9640f3049\" (UID: \"e3137730-712f-43ab-8c7c-bfd9640f3049\") " Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.582992 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e3137730-712f-43ab-8c7c-bfd9640f3049-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.584107 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e3137730-712f-43ab-8c7c-bfd9640f3049" (UID: "e3137730-712f-43ab-8c7c-bfd9640f3049"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.589149 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3137730-712f-43ab-8c7c-bfd9640f3049-kube-api-access-v645m" (OuterVolumeSpecName: "kube-api-access-v645m") pod "e3137730-712f-43ab-8c7c-bfd9640f3049" (UID: "e3137730-712f-43ab-8c7c-bfd9640f3049"). InnerVolumeSpecName "kube-api-access-v645m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.599989 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bgnql"] Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.607269 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-bgnql"] Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.614319 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/881ca87a-0f16-428d-b4e7-f9832b7b6f3e-kube-api-access-v96p9" (OuterVolumeSpecName: "kube-api-access-v96p9") pod "881ca87a-0f16-428d-b4e7-f9832b7b6f3e" (UID: "881ca87a-0f16-428d-b4e7-f9832b7b6f3e"). InnerVolumeSpecName "kube-api-access-v96p9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.635055 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-scripts" (OuterVolumeSpecName: "scripts") pod "e3137730-712f-43ab-8c7c-bfd9640f3049" (UID: "e3137730-712f-43ab-8c7c-bfd9640f3049"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.638708 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e3137730-712f-43ab-8c7c-bfd9640f3049" (UID: "e3137730-712f-43ab-8c7c-bfd9640f3049"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.640650 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e3137730-712f-43ab-8c7c-bfd9640f3049" (UID: "e3137730-712f-43ab-8c7c-bfd9640f3049"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.684617 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.684649 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v96p9\" (UniqueName: \"kubernetes.io/projected/881ca87a-0f16-428d-b4e7-f9832b7b6f3e-kube-api-access-v96p9\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.684663 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.684675 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e3137730-712f-43ab-8c7c-bfd9640f3049-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.684685 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v645m\" (UniqueName: \"kubernetes.io/projected/e3137730-712f-43ab-8c7c-bfd9640f3049-kube-api-access-v645m\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:03.684717 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e3137730-712f-43ab-8c7c-bfd9640f3049-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.082134 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3137730-712f-43ab-8c7c-bfd9640f3049" path="/var/lib/kubelet/pods/e3137730-712f-43ab-8c7c-bfd9640f3049/volumes" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.113426 4692 scope.go:117] "RemoveContainer" containerID="8033743fc6419ba0a01e3a0e5cac0d473add21dc8f4bcfc2aac21723d388ff61" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.113658 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-bgnql" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.120445 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" event={"ID":"881ca87a-0f16-428d-b4e7-f9832b7b6f3e","Type":"ContainerDied","Data":"d8c1a12d12371e72852817490872e4f0f9b0ab7589c2cc13436161e6fdd7443d"} Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.120497 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8c1a12d12371e72852817490872e4f0f9b0ab7589c2cc13436161e6fdd7443d" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.120529 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550838-9zpjw" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.587673 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550832-486cx"] Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.596521 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550832-486cx"] Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.623851 4692 scope.go:117] "RemoveContainer" containerID="21a1e11b7f3d4f058b58cb33009cee8f857f530015b70a25acc6fb0d7936bcc6" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.761529 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd"] Mar 09 09:58:04 crc kubenswrapper[4692]: E0309 09:58:04.763214 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881ca87a-0f16-428d-b4e7-f9832b7b6f3e" containerName="oc" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.763244 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="881ca87a-0f16-428d-b4e7-f9832b7b6f3e" containerName="oc" Mar 09 09:58:04 crc kubenswrapper[4692]: E0309 09:58:04.763363 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3137730-712f-43ab-8c7c-bfd9640f3049" containerName="swift-ring-rebalance" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.763393 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3137730-712f-43ab-8c7c-bfd9640f3049" containerName="swift-ring-rebalance" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.768121 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3137730-712f-43ab-8c7c-bfd9640f3049" containerName="swift-ring-rebalance" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.768210 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="881ca87a-0f16-428d-b4e7-f9832b7b6f3e" containerName="oc" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.769771 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.778815 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.778846 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.802553 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd"] Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.804180 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6mzd\" (UniqueName: \"kubernetes.io/projected/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-kube-api-access-g6mzd\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.804244 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-dispersionconf\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.804292 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-scripts\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.804381 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.804400 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-swiftconf\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.804731 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-etc-swift\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.906817 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-etc-swift\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.907356 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-etc-swift\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.906954 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6mzd\" (UniqueName: \"kubernetes.io/projected/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-kube-api-access-g6mzd\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.907706 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-dispersionconf\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.908845 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-scripts\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.909213 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-swiftconf\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.909250 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.909610 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-scripts\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.910088 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.916285 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-dispersionconf\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.916815 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-swiftconf\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:04 crc kubenswrapper[4692]: I0309 09:58:04.924830 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6mzd\" (UniqueName: \"kubernetes.io/projected/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-kube-api-access-g6mzd\") pod \"swift-ring-rebalance-debug-kwnkd\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:05 crc kubenswrapper[4692]: I0309 09:58:05.100821 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:05 crc kubenswrapper[4692]: I0309 09:58:05.563223 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd"] Mar 09 09:58:06 crc kubenswrapper[4692]: I0309 09:58:06.080260 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d10c9253-78bc-437f-8dd6-34830d6a263c" path="/var/lib/kubelet/pods/d10c9253-78bc-437f-8dd6-34830d6a263c/volumes" Mar 09 09:58:06 crc kubenswrapper[4692]: I0309 09:58:06.154446 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" event={"ID":"30756dc0-8fce-4129-9bad-1b2e3aa09f1d","Type":"ContainerStarted","Data":"c754d1774f15d29debcb96e5587bb6e5cbcb5bec8fade21eb539b209681a92b5"} Mar 09 09:58:06 crc kubenswrapper[4692]: I0309 09:58:06.154501 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" event={"ID":"30756dc0-8fce-4129-9bad-1b2e3aa09f1d","Type":"ContainerStarted","Data":"04b651ddbb000b59bc421d143bae79cc818395a6793696cd3a6de06177d1d8a9"} Mar 09 09:58:06 crc kubenswrapper[4692]: I0309 09:58:06.188669 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" podStartSLOduration=2.188645031 podStartE2EDuration="2.188645031s" podCreationTimestamp="2026-03-09 09:58:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:06.184800623 +0000 UTC m=+2287.009536214" watchObservedRunningTime="2026-03-09 09:58:06.188645031 +0000 UTC m=+2287.013380623" Mar 09 09:58:07 crc kubenswrapper[4692]: I0309 09:58:07.071650 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:58:07 crc kubenswrapper[4692]: E0309 09:58:07.072001 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:58:08 crc kubenswrapper[4692]: I0309 09:58:08.178450 4692 generic.go:334] "Generic (PLEG): container finished" podID="30756dc0-8fce-4129-9bad-1b2e3aa09f1d" containerID="c754d1774f15d29debcb96e5587bb6e5cbcb5bec8fade21eb539b209681a92b5" exitCode=0 Mar 09 09:58:08 crc kubenswrapper[4692]: I0309 09:58:08.178555 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" event={"ID":"30756dc0-8fce-4129-9bad-1b2e3aa09f1d","Type":"ContainerDied","Data":"c754d1774f15d29debcb96e5587bb6e5cbcb5bec8fade21eb539b209681a92b5"} Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.501852 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.535732 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd"] Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.545613 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd"] Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.592331 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-ring-data-devices\") pod \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.592415 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-dispersionconf\") pod \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.592477 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-swiftconf\") pod \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.592513 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-scripts\") pod \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.592602 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6mzd\" (UniqueName: \"kubernetes.io/projected/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-kube-api-access-g6mzd\") pod \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.592661 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-etc-swift\") pod \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\" (UID: \"30756dc0-8fce-4129-9bad-1b2e3aa09f1d\") " Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.593768 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "30756dc0-8fce-4129-9bad-1b2e3aa09f1d" (UID: "30756dc0-8fce-4129-9bad-1b2e3aa09f1d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.593910 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "30756dc0-8fce-4129-9bad-1b2e3aa09f1d" (UID: "30756dc0-8fce-4129-9bad-1b2e3aa09f1d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.610287 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-kube-api-access-g6mzd" (OuterVolumeSpecName: "kube-api-access-g6mzd") pod "30756dc0-8fce-4129-9bad-1b2e3aa09f1d" (UID: "30756dc0-8fce-4129-9bad-1b2e3aa09f1d"). InnerVolumeSpecName "kube-api-access-g6mzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.620066 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "30756dc0-8fce-4129-9bad-1b2e3aa09f1d" (UID: "30756dc0-8fce-4129-9bad-1b2e3aa09f1d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.625090 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "30756dc0-8fce-4129-9bad-1b2e3aa09f1d" (UID: "30756dc0-8fce-4129-9bad-1b2e3aa09f1d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.631498 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-scripts" (OuterVolumeSpecName: "scripts") pod "30756dc0-8fce-4129-9bad-1b2e3aa09f1d" (UID: "30756dc0-8fce-4129-9bad-1b2e3aa09f1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.694811 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.695248 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.695294 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6mzd\" (UniqueName: \"kubernetes.io/projected/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-kube-api-access-g6mzd\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.695310 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.695322 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:09 crc kubenswrapper[4692]: I0309 09:58:09.695334 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/30756dc0-8fce-4129-9bad-1b2e3aa09f1d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.080255 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30756dc0-8fce-4129-9bad-1b2e3aa09f1d" path="/var/lib/kubelet/pods/30756dc0-8fce-4129-9bad-1b2e3aa09f1d/volumes" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.196681 4692 scope.go:117] "RemoveContainer" containerID="c754d1774f15d29debcb96e5587bb6e5cbcb5bec8fade21eb539b209681a92b5" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.196741 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kwnkd" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.702626 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7lq74"] Mar 09 09:58:10 crc kubenswrapper[4692]: E0309 09:58:10.703128 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30756dc0-8fce-4129-9bad-1b2e3aa09f1d" containerName="swift-ring-rebalance" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.703148 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="30756dc0-8fce-4129-9bad-1b2e3aa09f1d" containerName="swift-ring-rebalance" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.703388 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="30756dc0-8fce-4129-9bad-1b2e3aa09f1d" containerName="swift-ring-rebalance" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.704076 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.706467 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.706656 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.713729 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7lq74"] Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.813714 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-dispersionconf\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.813981 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65zzr\" (UniqueName: \"kubernetes.io/projected/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-kube-api-access-65zzr\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.814086 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-ring-data-devices\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.814139 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-scripts\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.814186 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-swiftconf\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.814376 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-etc-swift\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.915534 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-etc-swift\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.915610 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-dispersionconf\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.915681 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65zzr\" (UniqueName: \"kubernetes.io/projected/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-kube-api-access-65zzr\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.915717 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-ring-data-devices\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.915746 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-scripts\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.915770 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-swiftconf\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.916148 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-etc-swift\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.916879 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-scripts\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.917253 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-ring-data-devices\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.921602 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-dispersionconf\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.921927 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-swiftconf\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:10 crc kubenswrapper[4692]: I0309 09:58:10.941191 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65zzr\" (UniqueName: \"kubernetes.io/projected/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-kube-api-access-65zzr\") pod \"swift-ring-rebalance-debug-7lq74\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:11 crc kubenswrapper[4692]: I0309 09:58:11.025272 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:11 crc kubenswrapper[4692]: I0309 09:58:11.468490 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7lq74"] Mar 09 09:58:12 crc kubenswrapper[4692]: I0309 09:58:12.244790 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" event={"ID":"11fd1278-d2bc-4f73-8138-2fd51e9f19d9","Type":"ContainerStarted","Data":"d4d672cf55b6bbc477e4745fbea789482e5f0eb874d6fe2f0acee870e9211993"} Mar 09 09:58:12 crc kubenswrapper[4692]: I0309 09:58:12.245139 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" event={"ID":"11fd1278-d2bc-4f73-8138-2fd51e9f19d9","Type":"ContainerStarted","Data":"be387678e413962489190933cedef14d9f7837d6d83af2e72e18a55aba5a34f1"} Mar 09 09:58:12 crc kubenswrapper[4692]: I0309 09:58:12.279810 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" podStartSLOduration=2.279790309 podStartE2EDuration="2.279790309s" podCreationTimestamp="2026-03-09 09:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:12.266316649 +0000 UTC m=+2293.091052230" watchObservedRunningTime="2026-03-09 09:58:12.279790309 +0000 UTC m=+2293.104525890" Mar 09 09:58:13 crc kubenswrapper[4692]: I0309 09:58:13.259559 4692 generic.go:334] "Generic (PLEG): container finished" podID="11fd1278-d2bc-4f73-8138-2fd51e9f19d9" containerID="d4d672cf55b6bbc477e4745fbea789482e5f0eb874d6fe2f0acee870e9211993" exitCode=0 Mar 09 09:58:13 crc kubenswrapper[4692]: I0309 09:58:13.259670 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" event={"ID":"11fd1278-d2bc-4f73-8138-2fd51e9f19d9","Type":"ContainerDied","Data":"d4d672cf55b6bbc477e4745fbea789482e5f0eb874d6fe2f0acee870e9211993"} Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.563202 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.597900 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7lq74"] Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.603274 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7lq74"] Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.686386 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-ring-data-devices\") pod \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.686561 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-swiftconf\") pod \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.686704 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-etc-swift\") pod \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.686801 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-dispersionconf\") pod \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.686834 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-scripts\") pod \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.686926 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65zzr\" (UniqueName: \"kubernetes.io/projected/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-kube-api-access-65zzr\") pod \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\" (UID: \"11fd1278-d2bc-4f73-8138-2fd51e9f19d9\") " Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.687675 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "11fd1278-d2bc-4f73-8138-2fd51e9f19d9" (UID: "11fd1278-d2bc-4f73-8138-2fd51e9f19d9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.687807 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "11fd1278-d2bc-4f73-8138-2fd51e9f19d9" (UID: "11fd1278-d2bc-4f73-8138-2fd51e9f19d9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.694215 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-kube-api-access-65zzr" (OuterVolumeSpecName: "kube-api-access-65zzr") pod "11fd1278-d2bc-4f73-8138-2fd51e9f19d9" (UID: "11fd1278-d2bc-4f73-8138-2fd51e9f19d9"). InnerVolumeSpecName "kube-api-access-65zzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.719741 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-scripts" (OuterVolumeSpecName: "scripts") pod "11fd1278-d2bc-4f73-8138-2fd51e9f19d9" (UID: "11fd1278-d2bc-4f73-8138-2fd51e9f19d9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.722187 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "11fd1278-d2bc-4f73-8138-2fd51e9f19d9" (UID: "11fd1278-d2bc-4f73-8138-2fd51e9f19d9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.726372 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "11fd1278-d2bc-4f73-8138-2fd51e9f19d9" (UID: "11fd1278-d2bc-4f73-8138-2fd51e9f19d9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.788936 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.788984 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.789000 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.789012 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65zzr\" (UniqueName: \"kubernetes.io/projected/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-kube-api-access-65zzr\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.789023 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:14 crc kubenswrapper[4692]: I0309 09:58:14.789034 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/11fd1278-d2bc-4f73-8138-2fd51e9f19d9-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.281512 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be387678e413962489190933cedef14d9f7837d6d83af2e72e18a55aba5a34f1" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.281571 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7lq74" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.739512 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf"] Mar 09 09:58:15 crc kubenswrapper[4692]: E0309 09:58:15.739850 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fd1278-d2bc-4f73-8138-2fd51e9f19d9" containerName="swift-ring-rebalance" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.739869 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fd1278-d2bc-4f73-8138-2fd51e9f19d9" containerName="swift-ring-rebalance" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.740059 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fd1278-d2bc-4f73-8138-2fd51e9f19d9" containerName="swift-ring-rebalance" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.740716 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.742807 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.743348 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.758150 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf"] Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.804427 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-scripts\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.804548 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-ring-data-devices\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.804621 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-dispersionconf\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.804702 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s75sd\" (UniqueName: \"kubernetes.io/projected/54f23b66-355f-4602-bed5-bda2c2b0e111-kube-api-access-s75sd\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.804767 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-swiftconf\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.804795 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/54f23b66-355f-4602-bed5-bda2c2b0e111-etc-swift\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.906574 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s75sd\" (UniqueName: \"kubernetes.io/projected/54f23b66-355f-4602-bed5-bda2c2b0e111-kube-api-access-s75sd\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.906674 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-swiftconf\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.906709 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/54f23b66-355f-4602-bed5-bda2c2b0e111-etc-swift\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.906810 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-scripts\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.906851 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-ring-data-devices\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.906881 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-dispersionconf\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.907828 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/54f23b66-355f-4602-bed5-bda2c2b0e111-etc-swift\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.908391 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-ring-data-devices\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.908891 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-scripts\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.913928 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-dispersionconf\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.915354 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-swiftconf\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:15 crc kubenswrapper[4692]: I0309 09:58:15.932979 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s75sd\" (UniqueName: \"kubernetes.io/projected/54f23b66-355f-4602-bed5-bda2c2b0e111-kube-api-access-s75sd\") pod \"swift-ring-rebalance-debug-2ntgf\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:16 crc kubenswrapper[4692]: I0309 09:58:16.069260 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:16 crc kubenswrapper[4692]: I0309 09:58:16.083712 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11fd1278-d2bc-4f73-8138-2fd51e9f19d9" path="/var/lib/kubelet/pods/11fd1278-d2bc-4f73-8138-2fd51e9f19d9/volumes" Mar 09 09:58:16 crc kubenswrapper[4692]: I0309 09:58:16.353473 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf"] Mar 09 09:58:16 crc kubenswrapper[4692]: W0309 09:58:16.364389 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54f23b66_355f_4602_bed5_bda2c2b0e111.slice/crio-17b9c9938d9b34f9607302bcfa3d0fcb46111a1b20affc2fc6fbc4d6cf73f76c WatchSource:0}: Error finding container 17b9c9938d9b34f9607302bcfa3d0fcb46111a1b20affc2fc6fbc4d6cf73f76c: Status 404 returned error can't find the container with id 17b9c9938d9b34f9607302bcfa3d0fcb46111a1b20affc2fc6fbc4d6cf73f76c Mar 09 09:58:17 crc kubenswrapper[4692]: I0309 09:58:17.317748 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" event={"ID":"54f23b66-355f-4602-bed5-bda2c2b0e111","Type":"ContainerStarted","Data":"3fc7300ce06cac6fa7ffd53b921ee4ecd80af7f9514c4d1e4994dda27352491f"} Mar 09 09:58:17 crc kubenswrapper[4692]: I0309 09:58:17.318056 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" event={"ID":"54f23b66-355f-4602-bed5-bda2c2b0e111","Type":"ContainerStarted","Data":"17b9c9938d9b34f9607302bcfa3d0fcb46111a1b20affc2fc6fbc4d6cf73f76c"} Mar 09 09:58:17 crc kubenswrapper[4692]: I0309 09:58:17.337921 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" podStartSLOduration=2.337906105 podStartE2EDuration="2.337906105s" podCreationTimestamp="2026-03-09 09:58:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:17.336778274 +0000 UTC m=+2298.161513855" watchObservedRunningTime="2026-03-09 09:58:17.337906105 +0000 UTC m=+2298.162641686" Mar 09 09:58:18 crc kubenswrapper[4692]: I0309 09:58:18.335372 4692 generic.go:334] "Generic (PLEG): container finished" podID="54f23b66-355f-4602-bed5-bda2c2b0e111" containerID="3fc7300ce06cac6fa7ffd53b921ee4ecd80af7f9514c4d1e4994dda27352491f" exitCode=0 Mar 09 09:58:18 crc kubenswrapper[4692]: I0309 09:58:18.335425 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" event={"ID":"54f23b66-355f-4602-bed5-bda2c2b0e111","Type":"ContainerDied","Data":"3fc7300ce06cac6fa7ffd53b921ee4ecd80af7f9514c4d1e4994dda27352491f"} Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.071518 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:58:19 crc kubenswrapper[4692]: E0309 09:58:19.072266 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.651290 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.699714 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf"] Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.705092 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf"] Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.773462 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s75sd\" (UniqueName: \"kubernetes.io/projected/54f23b66-355f-4602-bed5-bda2c2b0e111-kube-api-access-s75sd\") pod \"54f23b66-355f-4602-bed5-bda2c2b0e111\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.773999 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/54f23b66-355f-4602-bed5-bda2c2b0e111-etc-swift\") pod \"54f23b66-355f-4602-bed5-bda2c2b0e111\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.774082 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-dispersionconf\") pod \"54f23b66-355f-4602-bed5-bda2c2b0e111\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.774114 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-scripts\") pod \"54f23b66-355f-4602-bed5-bda2c2b0e111\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.774182 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-ring-data-devices\") pod \"54f23b66-355f-4602-bed5-bda2c2b0e111\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.774249 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-swiftconf\") pod \"54f23b66-355f-4602-bed5-bda2c2b0e111\" (UID: \"54f23b66-355f-4602-bed5-bda2c2b0e111\") " Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.774819 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54f23b66-355f-4602-bed5-bda2c2b0e111-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "54f23b66-355f-4602-bed5-bda2c2b0e111" (UID: "54f23b66-355f-4602-bed5-bda2c2b0e111"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.775510 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "54f23b66-355f-4602-bed5-bda2c2b0e111" (UID: "54f23b66-355f-4602-bed5-bda2c2b0e111"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.781374 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54f23b66-355f-4602-bed5-bda2c2b0e111-kube-api-access-s75sd" (OuterVolumeSpecName: "kube-api-access-s75sd") pod "54f23b66-355f-4602-bed5-bda2c2b0e111" (UID: "54f23b66-355f-4602-bed5-bda2c2b0e111"). InnerVolumeSpecName "kube-api-access-s75sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.798288 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-scripts" (OuterVolumeSpecName: "scripts") pod "54f23b66-355f-4602-bed5-bda2c2b0e111" (UID: "54f23b66-355f-4602-bed5-bda2c2b0e111"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.805145 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "54f23b66-355f-4602-bed5-bda2c2b0e111" (UID: "54f23b66-355f-4602-bed5-bda2c2b0e111"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.811129 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "54f23b66-355f-4602-bed5-bda2c2b0e111" (UID: "54f23b66-355f-4602-bed5-bda2c2b0e111"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.877036 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s75sd\" (UniqueName: \"kubernetes.io/projected/54f23b66-355f-4602-bed5-bda2c2b0e111-kube-api-access-s75sd\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.877077 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/54f23b66-355f-4602-bed5-bda2c2b0e111-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.877086 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.877097 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.877107 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/54f23b66-355f-4602-bed5-bda2c2b0e111-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:19 crc kubenswrapper[4692]: I0309 09:58:19.877117 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/54f23b66-355f-4602-bed5-bda2c2b0e111-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.082794 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54f23b66-355f-4602-bed5-bda2c2b0e111" path="/var/lib/kubelet/pods/54f23b66-355f-4602-bed5-bda2c2b0e111/volumes" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.358189 4692 scope.go:117] "RemoveContainer" containerID="3fc7300ce06cac6fa7ffd53b921ee4ecd80af7f9514c4d1e4994dda27352491f" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.358303 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-2ntgf" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.863370 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb"] Mar 09 09:58:20 crc kubenswrapper[4692]: E0309 09:58:20.863701 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54f23b66-355f-4602-bed5-bda2c2b0e111" containerName="swift-ring-rebalance" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.863717 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="54f23b66-355f-4602-bed5-bda2c2b0e111" containerName="swift-ring-rebalance" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.863861 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="54f23b66-355f-4602-bed5-bda2c2b0e111" containerName="swift-ring-rebalance" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.865020 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.867492 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.867793 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.882227 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb"] Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.893979 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wzpl\" (UniqueName: \"kubernetes.io/projected/33f53a6c-71bf-4161-b21d-0e8663ccb918-kube-api-access-5wzpl\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.894051 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-ring-data-devices\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.894074 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-scripts\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.894095 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/33f53a6c-71bf-4161-b21d-0e8663ccb918-etc-swift\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.894112 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-dispersionconf\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.894136 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-swiftconf\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.996385 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/33f53a6c-71bf-4161-b21d-0e8663ccb918-etc-swift\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.996462 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-dispersionconf\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.996506 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-swiftconf\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.996641 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wzpl\" (UniqueName: \"kubernetes.io/projected/33f53a6c-71bf-4161-b21d-0e8663ccb918-kube-api-access-5wzpl\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.996717 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-ring-data-devices\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.996744 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-scripts\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.997150 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/33f53a6c-71bf-4161-b21d-0e8663ccb918-etc-swift\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.997902 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-scripts\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:20 crc kubenswrapper[4692]: I0309 09:58:20.997930 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-ring-data-devices\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:21 crc kubenswrapper[4692]: I0309 09:58:21.003570 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-swiftconf\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:21 crc kubenswrapper[4692]: I0309 09:58:21.003579 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-dispersionconf\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:21 crc kubenswrapper[4692]: I0309 09:58:21.017352 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wzpl\" (UniqueName: \"kubernetes.io/projected/33f53a6c-71bf-4161-b21d-0e8663ccb918-kube-api-access-5wzpl\") pod \"swift-ring-rebalance-debug-7qvwb\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:21 crc kubenswrapper[4692]: I0309 09:58:21.188523 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:21 crc kubenswrapper[4692]: I0309 09:58:21.628725 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb"] Mar 09 09:58:22 crc kubenswrapper[4692]: I0309 09:58:22.380230 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" event={"ID":"33f53a6c-71bf-4161-b21d-0e8663ccb918","Type":"ContainerStarted","Data":"ebcc8354c1009007af26c0d9c3d6d8574fb0427f871aa435511b9319b399e4f4"} Mar 09 09:58:22 crc kubenswrapper[4692]: I0309 09:58:22.380277 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" event={"ID":"33f53a6c-71bf-4161-b21d-0e8663ccb918","Type":"ContainerStarted","Data":"3a8b05ca6815fb6bd9b60dfa5f2a43d9350398ab6e85886a7a560e429d10a432"} Mar 09 09:58:22 crc kubenswrapper[4692]: I0309 09:58:22.406381 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" podStartSLOduration=2.406361745 podStartE2EDuration="2.406361745s" podCreationTimestamp="2026-03-09 09:58:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:22.397502464 +0000 UTC m=+2303.222238065" watchObservedRunningTime="2026-03-09 09:58:22.406361745 +0000 UTC m=+2303.231097336" Mar 09 09:58:23 crc kubenswrapper[4692]: I0309 09:58:23.391965 4692 generic.go:334] "Generic (PLEG): container finished" podID="33f53a6c-71bf-4161-b21d-0e8663ccb918" containerID="ebcc8354c1009007af26c0d9c3d6d8574fb0427f871aa435511b9319b399e4f4" exitCode=0 Mar 09 09:58:23 crc kubenswrapper[4692]: I0309 09:58:23.392033 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" event={"ID":"33f53a6c-71bf-4161-b21d-0e8663ccb918","Type":"ContainerDied","Data":"ebcc8354c1009007af26c0d9c3d6d8574fb0427f871aa435511b9319b399e4f4"} Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.743377 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.758861 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/33f53a6c-71bf-4161-b21d-0e8663ccb918-etc-swift\") pod \"33f53a6c-71bf-4161-b21d-0e8663ccb918\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.758915 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-dispersionconf\") pod \"33f53a6c-71bf-4161-b21d-0e8663ccb918\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.759018 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-scripts\") pod \"33f53a6c-71bf-4161-b21d-0e8663ccb918\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.759372 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-ring-data-devices\") pod \"33f53a6c-71bf-4161-b21d-0e8663ccb918\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.759414 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wzpl\" (UniqueName: \"kubernetes.io/projected/33f53a6c-71bf-4161-b21d-0e8663ccb918-kube-api-access-5wzpl\") pod \"33f53a6c-71bf-4161-b21d-0e8663ccb918\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.759575 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-swiftconf\") pod \"33f53a6c-71bf-4161-b21d-0e8663ccb918\" (UID: \"33f53a6c-71bf-4161-b21d-0e8663ccb918\") " Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.760387 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "33f53a6c-71bf-4161-b21d-0e8663ccb918" (UID: "33f53a6c-71bf-4161-b21d-0e8663ccb918"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.762600 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33f53a6c-71bf-4161-b21d-0e8663ccb918-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "33f53a6c-71bf-4161-b21d-0e8663ccb918" (UID: "33f53a6c-71bf-4161-b21d-0e8663ccb918"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.768627 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33f53a6c-71bf-4161-b21d-0e8663ccb918-kube-api-access-5wzpl" (OuterVolumeSpecName: "kube-api-access-5wzpl") pod "33f53a6c-71bf-4161-b21d-0e8663ccb918" (UID: "33f53a6c-71bf-4161-b21d-0e8663ccb918"). InnerVolumeSpecName "kube-api-access-5wzpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.784528 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb"] Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.785646 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-scripts" (OuterVolumeSpecName: "scripts") pod "33f53a6c-71bf-4161-b21d-0e8663ccb918" (UID: "33f53a6c-71bf-4161-b21d-0e8663ccb918"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.791883 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb"] Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.797346 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "33f53a6c-71bf-4161-b21d-0e8663ccb918" (UID: "33f53a6c-71bf-4161-b21d-0e8663ccb918"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.799325 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "33f53a6c-71bf-4161-b21d-0e8663ccb918" (UID: "33f53a6c-71bf-4161-b21d-0e8663ccb918"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.862139 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.862200 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wzpl\" (UniqueName: \"kubernetes.io/projected/33f53a6c-71bf-4161-b21d-0e8663ccb918-kube-api-access-5wzpl\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.862212 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.862221 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/33f53a6c-71bf-4161-b21d-0e8663ccb918-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.862234 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/33f53a6c-71bf-4161-b21d-0e8663ccb918-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:24 crc kubenswrapper[4692]: I0309 09:58:24.862243 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/33f53a6c-71bf-4161-b21d-0e8663ccb918-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.426137 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a8b05ca6815fb6bd9b60dfa5f2a43d9350398ab6e85886a7a560e429d10a432" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.426317 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7qvwb" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.933643 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5pltg"] Mar 09 09:58:25 crc kubenswrapper[4692]: E0309 09:58:25.933941 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33f53a6c-71bf-4161-b21d-0e8663ccb918" containerName="swift-ring-rebalance" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.933954 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="33f53a6c-71bf-4161-b21d-0e8663ccb918" containerName="swift-ring-rebalance" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.934112 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="33f53a6c-71bf-4161-b21d-0e8663ccb918" containerName="swift-ring-rebalance" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.934617 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.937659 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.938213 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.950268 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5pltg"] Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.984896 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-etc-swift\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.984974 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw8l5\" (UniqueName: \"kubernetes.io/projected/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-kube-api-access-zw8l5\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.985254 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-ring-data-devices\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.985462 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-scripts\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.985539 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-dispersionconf\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:25 crc kubenswrapper[4692]: I0309 09:58:25.985629 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-swiftconf\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.082353 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33f53a6c-71bf-4161-b21d-0e8663ccb918" path="/var/lib/kubelet/pods/33f53a6c-71bf-4161-b21d-0e8663ccb918/volumes" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.087404 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw8l5\" (UniqueName: \"kubernetes.io/projected/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-kube-api-access-zw8l5\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.087469 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-ring-data-devices\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.087507 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-scripts\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.087530 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-dispersionconf\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.087566 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-swiftconf\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.087610 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-etc-swift\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.088073 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-etc-swift\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.089009 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-ring-data-devices\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.090495 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-scripts\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.097061 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-swiftconf\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.097661 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-dispersionconf\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.114047 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw8l5\" (UniqueName: \"kubernetes.io/projected/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-kube-api-access-zw8l5\") pod \"swift-ring-rebalance-debug-5pltg\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.252971 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:26 crc kubenswrapper[4692]: I0309 09:58:26.534065 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5pltg"] Mar 09 09:58:27 crc kubenswrapper[4692]: I0309 09:58:27.453418 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" event={"ID":"8d28d79f-d1ee-4771-82da-2fb4b5d4e474","Type":"ContainerStarted","Data":"0784e5578126e1f55c1c0b6aa09e57960129619b22e14e369fcba131e956b0a6"} Mar 09 09:58:27 crc kubenswrapper[4692]: I0309 09:58:27.453767 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" event={"ID":"8d28d79f-d1ee-4771-82da-2fb4b5d4e474","Type":"ContainerStarted","Data":"90fb753dc34ed265316fa93fd8e5243af42a4275c96d936ac8caaf2342c00ca8"} Mar 09 09:58:27 crc kubenswrapper[4692]: I0309 09:58:27.479137 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" podStartSLOduration=2.479101104 podStartE2EDuration="2.479101104s" podCreationTimestamp="2026-03-09 09:58:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:27.47259203 +0000 UTC m=+2308.297327611" watchObservedRunningTime="2026-03-09 09:58:27.479101104 +0000 UTC m=+2308.303836695" Mar 09 09:58:28 crc kubenswrapper[4692]: I0309 09:58:28.489334 4692 generic.go:334] "Generic (PLEG): container finished" podID="8d28d79f-d1ee-4771-82da-2fb4b5d4e474" containerID="0784e5578126e1f55c1c0b6aa09e57960129619b22e14e369fcba131e956b0a6" exitCode=0 Mar 09 09:58:28 crc kubenswrapper[4692]: I0309 09:58:28.489385 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" event={"ID":"8d28d79f-d1ee-4771-82da-2fb4b5d4e474","Type":"ContainerDied","Data":"0784e5578126e1f55c1c0b6aa09e57960129619b22e14e369fcba131e956b0a6"} Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.793937 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.833920 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5pltg"] Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.839273 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-5pltg"] Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.953957 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-swiftconf\") pod \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.954031 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw8l5\" (UniqueName: \"kubernetes.io/projected/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-kube-api-access-zw8l5\") pod \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.954069 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-dispersionconf\") pod \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.954125 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-scripts\") pod \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.954153 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-ring-data-devices\") pod \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.954209 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-etc-swift\") pod \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\" (UID: \"8d28d79f-d1ee-4771-82da-2fb4b5d4e474\") " Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.955358 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8d28d79f-d1ee-4771-82da-2fb4b5d4e474" (UID: "8d28d79f-d1ee-4771-82da-2fb4b5d4e474"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.955713 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8d28d79f-d1ee-4771-82da-2fb4b5d4e474" (UID: "8d28d79f-d1ee-4771-82da-2fb4b5d4e474"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.960597 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-kube-api-access-zw8l5" (OuterVolumeSpecName: "kube-api-access-zw8l5") pod "8d28d79f-d1ee-4771-82da-2fb4b5d4e474" (UID: "8d28d79f-d1ee-4771-82da-2fb4b5d4e474"). InnerVolumeSpecName "kube-api-access-zw8l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.980935 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-scripts" (OuterVolumeSpecName: "scripts") pod "8d28d79f-d1ee-4771-82da-2fb4b5d4e474" (UID: "8d28d79f-d1ee-4771-82da-2fb4b5d4e474"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.982082 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8d28d79f-d1ee-4771-82da-2fb4b5d4e474" (UID: "8d28d79f-d1ee-4771-82da-2fb4b5d4e474"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:29 crc kubenswrapper[4692]: I0309 09:58:29.987594 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8d28d79f-d1ee-4771-82da-2fb4b5d4e474" (UID: "8d28d79f-d1ee-4771-82da-2fb4b5d4e474"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.056455 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.056505 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw8l5\" (UniqueName: \"kubernetes.io/projected/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-kube-api-access-zw8l5\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.056519 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.056530 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.056540 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.056561 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8d28d79f-d1ee-4771-82da-2fb4b5d4e474-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.085382 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d28d79f-d1ee-4771-82da-2fb4b5d4e474" path="/var/lib/kubelet/pods/8d28d79f-d1ee-4771-82da-2fb4b5d4e474/volumes" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.510276 4692 scope.go:117] "RemoveContainer" containerID="0784e5578126e1f55c1c0b6aa09e57960129619b22e14e369fcba131e956b0a6" Mar 09 09:58:30 crc kubenswrapper[4692]: I0309 09:58:30.510359 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-5pltg" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.013118 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-766fz"] Mar 09 09:58:31 crc kubenswrapper[4692]: E0309 09:58:31.013535 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d28d79f-d1ee-4771-82da-2fb4b5d4e474" containerName="swift-ring-rebalance" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.013555 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d28d79f-d1ee-4771-82da-2fb4b5d4e474" containerName="swift-ring-rebalance" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.013724 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d28d79f-d1ee-4771-82da-2fb4b5d4e474" containerName="swift-ring-rebalance" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.014337 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.017872 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.018779 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.039789 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-766fz"] Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.073113 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-ring-data-devices\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.073250 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fb8c2c6-6392-4279-9603-e8b08b0c4114-etc-swift\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.073308 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-swiftconf\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.073371 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-dispersionconf\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.073431 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-scripts\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.073555 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrk4w\" (UniqueName: \"kubernetes.io/projected/5fb8c2c6-6392-4279-9603-e8b08b0c4114-kube-api-access-mrk4w\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.173988 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-ring-data-devices\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.174033 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fb8c2c6-6392-4279-9603-e8b08b0c4114-etc-swift\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.174056 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-swiftconf\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.174083 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-dispersionconf\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.174119 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-scripts\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.174185 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrk4w\" (UniqueName: \"kubernetes.io/projected/5fb8c2c6-6392-4279-9603-e8b08b0c4114-kube-api-access-mrk4w\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.174937 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fb8c2c6-6392-4279-9603-e8b08b0c4114-etc-swift\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.175005 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-ring-data-devices\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.175484 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-scripts\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.181139 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-dispersionconf\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.183639 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-swiftconf\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.193848 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrk4w\" (UniqueName: \"kubernetes.io/projected/5fb8c2c6-6392-4279-9603-e8b08b0c4114-kube-api-access-mrk4w\") pod \"swift-ring-rebalance-debug-766fz\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.338211 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:31 crc kubenswrapper[4692]: I0309 09:58:31.617706 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-766fz"] Mar 09 09:58:32 crc kubenswrapper[4692]: I0309 09:58:32.533008 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" event={"ID":"5fb8c2c6-6392-4279-9603-e8b08b0c4114","Type":"ContainerStarted","Data":"97f4b3321a2f2c55b07b999e3c38bc44f9dd9bb5923fd87a74d7c0245b9bf2db"} Mar 09 09:58:32 crc kubenswrapper[4692]: I0309 09:58:32.533332 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" event={"ID":"5fb8c2c6-6392-4279-9603-e8b08b0c4114","Type":"ContainerStarted","Data":"a4f230b9a327ad4e4dbbb2922163b6df04939b5e774a49606523014f94ba4f53"} Mar 09 09:58:32 crc kubenswrapper[4692]: I0309 09:58:32.553031 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" podStartSLOduration=2.553011568 podStartE2EDuration="2.553011568s" podCreationTimestamp="2026-03-09 09:58:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:32.548383647 +0000 UTC m=+2313.373119238" watchObservedRunningTime="2026-03-09 09:58:32.553011568 +0000 UTC m=+2313.377747149" Mar 09 09:58:33 crc kubenswrapper[4692]: I0309 09:58:33.543895 4692 generic.go:334] "Generic (PLEG): container finished" podID="5fb8c2c6-6392-4279-9603-e8b08b0c4114" containerID="97f4b3321a2f2c55b07b999e3c38bc44f9dd9bb5923fd87a74d7c0245b9bf2db" exitCode=0 Mar 09 09:58:33 crc kubenswrapper[4692]: I0309 09:58:33.544093 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" event={"ID":"5fb8c2c6-6392-4279-9603-e8b08b0c4114","Type":"ContainerDied","Data":"97f4b3321a2f2c55b07b999e3c38bc44f9dd9bb5923fd87a74d7c0245b9bf2db"} Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.074498 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:58:34 crc kubenswrapper[4692]: E0309 09:58:34.074828 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.841783 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.889420 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-766fz"] Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.895563 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-766fz"] Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.935863 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-dispersionconf\") pod \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.935928 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-ring-data-devices\") pod \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.936047 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-swiftconf\") pod \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.936144 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-scripts\") pod \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.936269 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fb8c2c6-6392-4279-9603-e8b08b0c4114-etc-swift\") pod \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.936391 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrk4w\" (UniqueName: \"kubernetes.io/projected/5fb8c2c6-6392-4279-9603-e8b08b0c4114-kube-api-access-mrk4w\") pod \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\" (UID: \"5fb8c2c6-6392-4279-9603-e8b08b0c4114\") " Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.937103 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5fb8c2c6-6392-4279-9603-e8b08b0c4114" (UID: "5fb8c2c6-6392-4279-9603-e8b08b0c4114"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.937693 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fb8c2c6-6392-4279-9603-e8b08b0c4114-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5fb8c2c6-6392-4279-9603-e8b08b0c4114" (UID: "5fb8c2c6-6392-4279-9603-e8b08b0c4114"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.943241 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fb8c2c6-6392-4279-9603-e8b08b0c4114-kube-api-access-mrk4w" (OuterVolumeSpecName: "kube-api-access-mrk4w") pod "5fb8c2c6-6392-4279-9603-e8b08b0c4114" (UID: "5fb8c2c6-6392-4279-9603-e8b08b0c4114"). InnerVolumeSpecName "kube-api-access-mrk4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.958523 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-scripts" (OuterVolumeSpecName: "scripts") pod "5fb8c2c6-6392-4279-9603-e8b08b0c4114" (UID: "5fb8c2c6-6392-4279-9603-e8b08b0c4114"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.960603 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5fb8c2c6-6392-4279-9603-e8b08b0c4114" (UID: "5fb8c2c6-6392-4279-9603-e8b08b0c4114"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:34 crc kubenswrapper[4692]: I0309 09:58:34.962191 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5fb8c2c6-6392-4279-9603-e8b08b0c4114" (UID: "5fb8c2c6-6392-4279-9603-e8b08b0c4114"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.038658 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrk4w\" (UniqueName: \"kubernetes.io/projected/5fb8c2c6-6392-4279-9603-e8b08b0c4114-kube-api-access-mrk4w\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.038707 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.038722 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.038739 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5fb8c2c6-6392-4279-9603-e8b08b0c4114-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.038753 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5fb8c2c6-6392-4279-9603-e8b08b0c4114-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.038766 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5fb8c2c6-6392-4279-9603-e8b08b0c4114-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.570584 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4f230b9a327ad4e4dbbb2922163b6df04939b5e774a49606523014f94ba4f53" Mar 09 09:58:35 crc kubenswrapper[4692]: I0309 09:58:35.570636 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-766fz" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.034670 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jr46j"] Mar 09 09:58:36 crc kubenswrapper[4692]: E0309 09:58:36.035090 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fb8c2c6-6392-4279-9603-e8b08b0c4114" containerName="swift-ring-rebalance" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.035104 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fb8c2c6-6392-4279-9603-e8b08b0c4114" containerName="swift-ring-rebalance" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.035320 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fb8c2c6-6392-4279-9603-e8b08b0c4114" containerName="swift-ring-rebalance" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.035970 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.038491 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.041718 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.048566 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jr46j"] Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.089228 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-ring-data-devices\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.089321 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-swiftconf\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.089448 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-scripts\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.089511 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f2kg\" (UniqueName: \"kubernetes.io/projected/782e57e6-d939-4da7-b8d2-d910dedb50e8-kube-api-access-6f2kg\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.089768 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-dispersionconf\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.089888 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/782e57e6-d939-4da7-b8d2-d910dedb50e8-etc-swift\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.110835 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fb8c2c6-6392-4279-9603-e8b08b0c4114" path="/var/lib/kubelet/pods/5fb8c2c6-6392-4279-9603-e8b08b0c4114/volumes" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.193152 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-dispersionconf\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.195055 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/782e57e6-d939-4da7-b8d2-d910dedb50e8-etc-swift\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.195581 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-ring-data-devices\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.195696 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-swiftconf\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.195789 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-scripts\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.195905 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f2kg\" (UniqueName: \"kubernetes.io/projected/782e57e6-d939-4da7-b8d2-d910dedb50e8-kube-api-access-6f2kg\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.195520 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/782e57e6-d939-4da7-b8d2-d910dedb50e8-etc-swift\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.197356 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-ring-data-devices\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.199795 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-scripts\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.201718 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-dispersionconf\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.221974 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f2kg\" (UniqueName: \"kubernetes.io/projected/782e57e6-d939-4da7-b8d2-d910dedb50e8-kube-api-access-6f2kg\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.222027 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-swiftconf\") pod \"swift-ring-rebalance-debug-jr46j\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.431338 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:36 crc kubenswrapper[4692]: I0309 09:58:36.839097 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jr46j"] Mar 09 09:58:37 crc kubenswrapper[4692]: I0309 09:58:37.590720 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" event={"ID":"782e57e6-d939-4da7-b8d2-d910dedb50e8","Type":"ContainerStarted","Data":"6fca884ead2b04eee394bee0b7bda8fdb7c4edcb5cb1606de995b779f4ad7e87"} Mar 09 09:58:37 crc kubenswrapper[4692]: I0309 09:58:37.591300 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" event={"ID":"782e57e6-d939-4da7-b8d2-d910dedb50e8","Type":"ContainerStarted","Data":"52172e553b62ec0ae8899bfe6a663e9f96fddd81e45b0b9479e7b86da8c02e66"} Mar 09 09:58:37 crc kubenswrapper[4692]: I0309 09:58:37.610900 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" podStartSLOduration=1.6108797689999999 podStartE2EDuration="1.610879769s" podCreationTimestamp="2026-03-09 09:58:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:37.604464948 +0000 UTC m=+2318.429200529" watchObservedRunningTime="2026-03-09 09:58:37.610879769 +0000 UTC m=+2318.435615350" Mar 09 09:58:38 crc kubenswrapper[4692]: I0309 09:58:38.603025 4692 generic.go:334] "Generic (PLEG): container finished" podID="782e57e6-d939-4da7-b8d2-d910dedb50e8" containerID="6fca884ead2b04eee394bee0b7bda8fdb7c4edcb5cb1606de995b779f4ad7e87" exitCode=0 Mar 09 09:58:38 crc kubenswrapper[4692]: I0309 09:58:38.603185 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" event={"ID":"782e57e6-d939-4da7-b8d2-d910dedb50e8","Type":"ContainerDied","Data":"6fca884ead2b04eee394bee0b7bda8fdb7c4edcb5cb1606de995b779f4ad7e87"} Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.912781 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.953641 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jr46j"] Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.955021 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-ring-data-devices\") pod \"782e57e6-d939-4da7-b8d2-d910dedb50e8\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.955067 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-swiftconf\") pod \"782e57e6-d939-4da7-b8d2-d910dedb50e8\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.955144 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-dispersionconf\") pod \"782e57e6-d939-4da7-b8d2-d910dedb50e8\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.955193 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-scripts\") pod \"782e57e6-d939-4da7-b8d2-d910dedb50e8\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.955260 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/782e57e6-d939-4da7-b8d2-d910dedb50e8-etc-swift\") pod \"782e57e6-d939-4da7-b8d2-d910dedb50e8\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.955295 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f2kg\" (UniqueName: \"kubernetes.io/projected/782e57e6-d939-4da7-b8d2-d910dedb50e8-kube-api-access-6f2kg\") pod \"782e57e6-d939-4da7-b8d2-d910dedb50e8\" (UID: \"782e57e6-d939-4da7-b8d2-d910dedb50e8\") " Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.956284 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/782e57e6-d939-4da7-b8d2-d910dedb50e8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "782e57e6-d939-4da7-b8d2-d910dedb50e8" (UID: "782e57e6-d939-4da7-b8d2-d910dedb50e8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.956283 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "782e57e6-d939-4da7-b8d2-d910dedb50e8" (UID: "782e57e6-d939-4da7-b8d2-d910dedb50e8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.956869 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/782e57e6-d939-4da7-b8d2-d910dedb50e8-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.956889 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.961258 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782e57e6-d939-4da7-b8d2-d910dedb50e8-kube-api-access-6f2kg" (OuterVolumeSpecName: "kube-api-access-6f2kg") pod "782e57e6-d939-4da7-b8d2-d910dedb50e8" (UID: "782e57e6-d939-4da7-b8d2-d910dedb50e8"). InnerVolumeSpecName "kube-api-access-6f2kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.965879 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-jr46j"] Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.977988 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-scripts" (OuterVolumeSpecName: "scripts") pod "782e57e6-d939-4da7-b8d2-d910dedb50e8" (UID: "782e57e6-d939-4da7-b8d2-d910dedb50e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.983149 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "782e57e6-d939-4da7-b8d2-d910dedb50e8" (UID: "782e57e6-d939-4da7-b8d2-d910dedb50e8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:39 crc kubenswrapper[4692]: I0309 09:58:39.987659 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "782e57e6-d939-4da7-b8d2-d910dedb50e8" (UID: "782e57e6-d939-4da7-b8d2-d910dedb50e8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:40 crc kubenswrapper[4692]: I0309 09:58:40.058995 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f2kg\" (UniqueName: \"kubernetes.io/projected/782e57e6-d939-4da7-b8d2-d910dedb50e8-kube-api-access-6f2kg\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:40 crc kubenswrapper[4692]: I0309 09:58:40.059618 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:40 crc kubenswrapper[4692]: I0309 09:58:40.059640 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/782e57e6-d939-4da7-b8d2-d910dedb50e8-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:40 crc kubenswrapper[4692]: I0309 09:58:40.059655 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/782e57e6-d939-4da7-b8d2-d910dedb50e8-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:40 crc kubenswrapper[4692]: I0309 09:58:40.082260 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782e57e6-d939-4da7-b8d2-d910dedb50e8" path="/var/lib/kubelet/pods/782e57e6-d939-4da7-b8d2-d910dedb50e8/volumes" Mar 09 09:58:40 crc kubenswrapper[4692]: I0309 09:58:40.623641 4692 scope.go:117] "RemoveContainer" containerID="6fca884ead2b04eee394bee0b7bda8fdb7c4edcb5cb1606de995b779f4ad7e87" Mar 09 09:58:40 crc kubenswrapper[4692]: I0309 09:58:40.623701 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-jr46j" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.100141 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf"] Mar 09 09:58:41 crc kubenswrapper[4692]: E0309 09:58:41.100582 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782e57e6-d939-4da7-b8d2-d910dedb50e8" containerName="swift-ring-rebalance" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.100601 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="782e57e6-d939-4da7-b8d2-d910dedb50e8" containerName="swift-ring-rebalance" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.100756 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="782e57e6-d939-4da7-b8d2-d910dedb50e8" containerName="swift-ring-rebalance" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.101267 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.103758 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.103861 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.113075 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf"] Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.178555 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-scripts\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.178673 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5x9d\" (UniqueName: \"kubernetes.io/projected/15ed889b-c875-43b9-bf92-0bba841212b3-kube-api-access-v5x9d\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.178720 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-swiftconf\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.178768 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-ring-data-devices\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.178807 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15ed889b-c875-43b9-bf92-0bba841212b3-etc-swift\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.178826 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-dispersionconf\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.280708 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5x9d\" (UniqueName: \"kubernetes.io/projected/15ed889b-c875-43b9-bf92-0bba841212b3-kube-api-access-v5x9d\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.280766 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-swiftconf\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.280799 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-ring-data-devices\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.280831 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15ed889b-c875-43b9-bf92-0bba841212b3-etc-swift\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.280855 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-dispersionconf\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.280938 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-scripts\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.281568 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15ed889b-c875-43b9-bf92-0bba841212b3-etc-swift\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.281876 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-ring-data-devices\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.281962 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-scripts\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.285469 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-dispersionconf\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.295009 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-swiftconf\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.297294 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5x9d\" (UniqueName: \"kubernetes.io/projected/15ed889b-c875-43b9-bf92-0bba841212b3-kube-api-access-v5x9d\") pod \"swift-ring-rebalance-debug-vf5lf\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.424878 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:41 crc kubenswrapper[4692]: I0309 09:58:41.816304 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf"] Mar 09 09:58:42 crc kubenswrapper[4692]: I0309 09:58:42.645330 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" event={"ID":"15ed889b-c875-43b9-bf92-0bba841212b3","Type":"ContainerStarted","Data":"9fcd1d22bd9e043bfb4a986f66b936f27793554af5d6483b965bfff394a1e29d"} Mar 09 09:58:42 crc kubenswrapper[4692]: I0309 09:58:42.645724 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" event={"ID":"15ed889b-c875-43b9-bf92-0bba841212b3","Type":"ContainerStarted","Data":"10b09e21fcb9c0d0a7744129c46783d2f3a01a5ea87ae9390835b696b62b2179"} Mar 09 09:58:42 crc kubenswrapper[4692]: I0309 09:58:42.671035 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" podStartSLOduration=1.671010903 podStartE2EDuration="1.671010903s" podCreationTimestamp="2026-03-09 09:58:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:42.665346722 +0000 UTC m=+2323.490082343" watchObservedRunningTime="2026-03-09 09:58:42.671010903 +0000 UTC m=+2323.495746494" Mar 09 09:58:43 crc kubenswrapper[4692]: I0309 09:58:43.656996 4692 generic.go:334] "Generic (PLEG): container finished" podID="15ed889b-c875-43b9-bf92-0bba841212b3" containerID="9fcd1d22bd9e043bfb4a986f66b936f27793554af5d6483b965bfff394a1e29d" exitCode=0 Mar 09 09:58:43 crc kubenswrapper[4692]: I0309 09:58:43.657042 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" event={"ID":"15ed889b-c875-43b9-bf92-0bba841212b3","Type":"ContainerDied","Data":"9fcd1d22bd9e043bfb4a986f66b936f27793554af5d6483b965bfff394a1e29d"} Mar 09 09:58:44 crc kubenswrapper[4692]: I0309 09:58:44.959047 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:44 crc kubenswrapper[4692]: I0309 09:58:44.987867 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf"] Mar 09 09:58:44 crc kubenswrapper[4692]: I0309 09:58:44.995043 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf"] Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.031063 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15ed889b-c875-43b9-bf92-0bba841212b3-etc-swift\") pod \"15ed889b-c875-43b9-bf92-0bba841212b3\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.031125 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-ring-data-devices\") pod \"15ed889b-c875-43b9-bf92-0bba841212b3\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.031259 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-dispersionconf\") pod \"15ed889b-c875-43b9-bf92-0bba841212b3\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.031290 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5x9d\" (UniqueName: \"kubernetes.io/projected/15ed889b-c875-43b9-bf92-0bba841212b3-kube-api-access-v5x9d\") pod \"15ed889b-c875-43b9-bf92-0bba841212b3\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.031322 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-swiftconf\") pod \"15ed889b-c875-43b9-bf92-0bba841212b3\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.031352 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-scripts\") pod \"15ed889b-c875-43b9-bf92-0bba841212b3\" (UID: \"15ed889b-c875-43b9-bf92-0bba841212b3\") " Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.031805 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ed889b-c875-43b9-bf92-0bba841212b3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "15ed889b-c875-43b9-bf92-0bba841212b3" (UID: "15ed889b-c875-43b9-bf92-0bba841212b3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.032607 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "15ed889b-c875-43b9-bf92-0bba841212b3" (UID: "15ed889b-c875-43b9-bf92-0bba841212b3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.037373 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15ed889b-c875-43b9-bf92-0bba841212b3-kube-api-access-v5x9d" (OuterVolumeSpecName: "kube-api-access-v5x9d") pod "15ed889b-c875-43b9-bf92-0bba841212b3" (UID: "15ed889b-c875-43b9-bf92-0bba841212b3"). InnerVolumeSpecName "kube-api-access-v5x9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.054144 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-scripts" (OuterVolumeSpecName: "scripts") pod "15ed889b-c875-43b9-bf92-0bba841212b3" (UID: "15ed889b-c875-43b9-bf92-0bba841212b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.062189 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "15ed889b-c875-43b9-bf92-0bba841212b3" (UID: "15ed889b-c875-43b9-bf92-0bba841212b3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.064087 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "15ed889b-c875-43b9-bf92-0bba841212b3" (UID: "15ed889b-c875-43b9-bf92-0bba841212b3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.071425 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:58:45 crc kubenswrapper[4692]: E0309 09:58:45.071737 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.133372 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15ed889b-c875-43b9-bf92-0bba841212b3-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.133413 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.133428 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.133441 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5x9d\" (UniqueName: \"kubernetes.io/projected/15ed889b-c875-43b9-bf92-0bba841212b3-kube-api-access-v5x9d\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.133454 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15ed889b-c875-43b9-bf92-0bba841212b3-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.133466 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15ed889b-c875-43b9-bf92-0bba841212b3-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.676042 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10b09e21fcb9c0d0a7744129c46783d2f3a01a5ea87ae9390835b696b62b2179" Mar 09 09:58:45 crc kubenswrapper[4692]: I0309 09:58:45.676122 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vf5lf" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.080780 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15ed889b-c875-43b9-bf92-0bba841212b3" path="/var/lib/kubelet/pods/15ed889b-c875-43b9-bf92-0bba841212b3/volumes" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.130913 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq"] Mar 09 09:58:46 crc kubenswrapper[4692]: E0309 09:58:46.131242 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ed889b-c875-43b9-bf92-0bba841212b3" containerName="swift-ring-rebalance" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.131291 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ed889b-c875-43b9-bf92-0bba841212b3" containerName="swift-ring-rebalance" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.131438 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ed889b-c875-43b9-bf92-0bba841212b3" containerName="swift-ring-rebalance" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.131949 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.134220 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.134315 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.140525 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq"] Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.149445 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqncn\" (UniqueName: \"kubernetes.io/projected/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-kube-api-access-sqncn\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.149548 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-dispersionconf\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.149583 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-etc-swift\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.149627 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-swiftconf\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.149659 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-scripts\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.149803 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-ring-data-devices\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.251366 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-ring-data-devices\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.251488 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqncn\" (UniqueName: \"kubernetes.io/projected/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-kube-api-access-sqncn\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.251524 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-dispersionconf\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.251550 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-etc-swift\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.251576 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-swiftconf\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.251596 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-scripts\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.252305 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-ring-data-devices\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.252533 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-etc-swift\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.252644 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-scripts\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.255146 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-dispersionconf\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.264248 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-swiftconf\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.266690 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqncn\" (UniqueName: \"kubernetes.io/projected/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-kube-api-access-sqncn\") pod \"swift-ring-rebalance-debug-ltkxq\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.452116 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:46 crc kubenswrapper[4692]: I0309 09:58:46.858448 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq"] Mar 09 09:58:47 crc kubenswrapper[4692]: I0309 09:58:47.694937 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" event={"ID":"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5","Type":"ContainerStarted","Data":"589cc4f0a066ba0c299651050033de51e68e2be669e97613b888ff4b74702a46"} Mar 09 09:58:47 crc kubenswrapper[4692]: I0309 09:58:47.694986 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" event={"ID":"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5","Type":"ContainerStarted","Data":"36af225eec6dbf0e414c33add490d304619bba2e2fdd88754506ab6d25d05a66"} Mar 09 09:58:47 crc kubenswrapper[4692]: I0309 09:58:47.714578 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" podStartSLOduration=1.7145600779999999 podStartE2EDuration="1.714560078s" podCreationTimestamp="2026-03-09 09:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:47.713901049 +0000 UTC m=+2328.538636650" watchObservedRunningTime="2026-03-09 09:58:47.714560078 +0000 UTC m=+2328.539295659" Mar 09 09:58:48 crc kubenswrapper[4692]: I0309 09:58:48.703879 4692 generic.go:334] "Generic (PLEG): container finished" podID="ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" containerID="589cc4f0a066ba0c299651050033de51e68e2be669e97613b888ff4b74702a46" exitCode=0 Mar 09 09:58:48 crc kubenswrapper[4692]: I0309 09:58:48.703939 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" event={"ID":"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5","Type":"ContainerDied","Data":"589cc4f0a066ba0c299651050033de51e68e2be669e97613b888ff4b74702a46"} Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.057433 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.102853 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq"] Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.110502 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq"] Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.128316 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-swiftconf\") pod \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.128385 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-etc-swift\") pod \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.128427 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-dispersionconf\") pod \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.128498 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqncn\" (UniqueName: \"kubernetes.io/projected/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-kube-api-access-sqncn\") pod \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.128544 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-ring-data-devices\") pod \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.128610 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-scripts\") pod \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\" (UID: \"ecb611f6-e342-44c9-8d78-71eb7c9f9aa5\") " Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.131205 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" (UID: "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.132281 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" (UID: "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.139424 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-kube-api-access-sqncn" (OuterVolumeSpecName: "kube-api-access-sqncn") pod "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" (UID: "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5"). InnerVolumeSpecName "kube-api-access-sqncn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.156635 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-scripts" (OuterVolumeSpecName: "scripts") pod "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" (UID: "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.165284 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" (UID: "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.177131 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" (UID: "ecb611f6-e342-44c9-8d78-71eb7c9f9aa5"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.230805 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.230846 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.230857 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqncn\" (UniqueName: \"kubernetes.io/projected/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-kube-api-access-sqncn\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.230867 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.230877 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.230885 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.730816 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36af225eec6dbf0e414c33add490d304619bba2e2fdd88754506ab6d25d05a66" Mar 09 09:58:50 crc kubenswrapper[4692]: I0309 09:58:50.731127 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-ltkxq" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.267719 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz"] Mar 09 09:58:51 crc kubenswrapper[4692]: E0309 09:58:51.268731 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" containerName="swift-ring-rebalance" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.268820 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" containerName="swift-ring-rebalance" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.269082 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" containerName="swift-ring-rebalance" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.269816 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.273406 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.277566 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.282201 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz"] Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.349330 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-swiftconf\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.349438 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f089d5f7-6340-419c-8ae4-640b1639a6e0-etc-swift\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.349537 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-ring-data-devices\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.349600 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-scripts\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.349637 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvmw8\" (UniqueName: \"kubernetes.io/projected/f089d5f7-6340-419c-8ae4-640b1639a6e0-kube-api-access-cvmw8\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.349670 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-dispersionconf\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.451726 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-swiftconf\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.451827 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f089d5f7-6340-419c-8ae4-640b1639a6e0-etc-swift\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.451907 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-ring-data-devices\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.451948 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-scripts\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.451977 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvmw8\" (UniqueName: \"kubernetes.io/projected/f089d5f7-6340-419c-8ae4-640b1639a6e0-kube-api-access-cvmw8\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.452010 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-dispersionconf\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.453590 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-ring-data-devices\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.453973 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f089d5f7-6340-419c-8ae4-640b1639a6e0-etc-swift\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.454176 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-scripts\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.457282 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-swiftconf\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.457643 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-dispersionconf\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.482305 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvmw8\" (UniqueName: \"kubernetes.io/projected/f089d5f7-6340-419c-8ae4-640b1639a6e0-kube-api-access-cvmw8\") pod \"swift-ring-rebalance-debug-vzkqz\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:51 crc kubenswrapper[4692]: I0309 09:58:51.590758 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:52 crc kubenswrapper[4692]: I0309 09:58:52.030522 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz"] Mar 09 09:58:52 crc kubenswrapper[4692]: I0309 09:58:52.087551 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb611f6-e342-44c9-8d78-71eb7c9f9aa5" path="/var/lib/kubelet/pods/ecb611f6-e342-44c9-8d78-71eb7c9f9aa5/volumes" Mar 09 09:58:52 crc kubenswrapper[4692]: I0309 09:58:52.758454 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" event={"ID":"f089d5f7-6340-419c-8ae4-640b1639a6e0","Type":"ContainerStarted","Data":"688051e31d1d9865ebe410f5447f2518d7bd92a0296c388ba4d03d8073eb57ab"} Mar 09 09:58:52 crc kubenswrapper[4692]: I0309 09:58:52.758529 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" event={"ID":"f089d5f7-6340-419c-8ae4-640b1639a6e0","Type":"ContainerStarted","Data":"9d44e93f1f5c7df9b08c2321699770e5a6c59c6b73f3da409ea9b4fe3f05e52b"} Mar 09 09:58:52 crc kubenswrapper[4692]: I0309 09:58:52.788103 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" podStartSLOduration=1.788083779 podStartE2EDuration="1.788083779s" podCreationTimestamp="2026-03-09 09:58:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:52.783257343 +0000 UTC m=+2333.607992944" watchObservedRunningTime="2026-03-09 09:58:52.788083779 +0000 UTC m=+2333.612819350" Mar 09 09:58:53 crc kubenswrapper[4692]: I0309 09:58:53.769736 4692 generic.go:334] "Generic (PLEG): container finished" podID="f089d5f7-6340-419c-8ae4-640b1639a6e0" containerID="688051e31d1d9865ebe410f5447f2518d7bd92a0296c388ba4d03d8073eb57ab" exitCode=0 Mar 09 09:58:53 crc kubenswrapper[4692]: I0309 09:58:53.769829 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" event={"ID":"f089d5f7-6340-419c-8ae4-640b1639a6e0","Type":"ContainerDied","Data":"688051e31d1d9865ebe410f5447f2518d7bd92a0296c388ba4d03d8073eb57ab"} Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.045114 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.072893 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz"] Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.081342 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz"] Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.111743 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvmw8\" (UniqueName: \"kubernetes.io/projected/f089d5f7-6340-419c-8ae4-640b1639a6e0-kube-api-access-cvmw8\") pod \"f089d5f7-6340-419c-8ae4-640b1639a6e0\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.111919 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-dispersionconf\") pod \"f089d5f7-6340-419c-8ae4-640b1639a6e0\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.111951 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-ring-data-devices\") pod \"f089d5f7-6340-419c-8ae4-640b1639a6e0\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.111993 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f089d5f7-6340-419c-8ae4-640b1639a6e0-etc-swift\") pod \"f089d5f7-6340-419c-8ae4-640b1639a6e0\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.112034 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-swiftconf\") pod \"f089d5f7-6340-419c-8ae4-640b1639a6e0\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.112063 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-scripts\") pod \"f089d5f7-6340-419c-8ae4-640b1639a6e0\" (UID: \"f089d5f7-6340-419c-8ae4-640b1639a6e0\") " Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.112578 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f089d5f7-6340-419c-8ae4-640b1639a6e0" (UID: "f089d5f7-6340-419c-8ae4-640b1639a6e0"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.113278 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f089d5f7-6340-419c-8ae4-640b1639a6e0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f089d5f7-6340-419c-8ae4-640b1639a6e0" (UID: "f089d5f7-6340-419c-8ae4-640b1639a6e0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.120416 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f089d5f7-6340-419c-8ae4-640b1639a6e0-kube-api-access-cvmw8" (OuterVolumeSpecName: "kube-api-access-cvmw8") pod "f089d5f7-6340-419c-8ae4-640b1639a6e0" (UID: "f089d5f7-6340-419c-8ae4-640b1639a6e0"). InnerVolumeSpecName "kube-api-access-cvmw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.130670 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-scripts" (OuterVolumeSpecName: "scripts") pod "f089d5f7-6340-419c-8ae4-640b1639a6e0" (UID: "f089d5f7-6340-419c-8ae4-640b1639a6e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.135947 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f089d5f7-6340-419c-8ae4-640b1639a6e0" (UID: "f089d5f7-6340-419c-8ae4-640b1639a6e0"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.136660 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f089d5f7-6340-419c-8ae4-640b1639a6e0" (UID: "f089d5f7-6340-419c-8ae4-640b1639a6e0"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.216040 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.216073 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.216081 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f089d5f7-6340-419c-8ae4-640b1639a6e0-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.216090 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f089d5f7-6340-419c-8ae4-640b1639a6e0-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.216098 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f089d5f7-6340-419c-8ae4-640b1639a6e0-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.216107 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvmw8\" (UniqueName: \"kubernetes.io/projected/f089d5f7-6340-419c-8ae4-640b1639a6e0-kube-api-access-cvmw8\") on node \"crc\" DevicePath \"\"" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.805421 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d44e93f1f5c7df9b08c2321699770e5a6c59c6b73f3da409ea9b4fe3f05e52b" Mar 09 09:58:55 crc kubenswrapper[4692]: I0309 09:58:55.805616 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vzkqz" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.080869 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f089d5f7-6340-419c-8ae4-640b1639a6e0" path="/var/lib/kubelet/pods/f089d5f7-6340-419c-8ae4-640b1639a6e0/volumes" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.222887 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-np59w"] Mar 09 09:58:56 crc kubenswrapper[4692]: E0309 09:58:56.223217 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f089d5f7-6340-419c-8ae4-640b1639a6e0" containerName="swift-ring-rebalance" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.223233 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="f089d5f7-6340-419c-8ae4-640b1639a6e0" containerName="swift-ring-rebalance" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.223439 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="f089d5f7-6340-419c-8ae4-640b1639a6e0" containerName="swift-ring-rebalance" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.223908 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.228441 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.228961 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.234761 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-np59w"] Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.331267 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7tk\" (UniqueName: \"kubernetes.io/projected/04810067-c268-4cae-905c-bb4913ae908e-kube-api-access-zx7tk\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.331338 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-ring-data-devices\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.331432 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-swiftconf\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.331494 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-scripts\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.331752 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-dispersionconf\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.331805 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/04810067-c268-4cae-905c-bb4913ae908e-etc-swift\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.433383 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-dispersionconf\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.433724 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/04810067-c268-4cae-905c-bb4913ae908e-etc-swift\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.433782 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7tk\" (UniqueName: \"kubernetes.io/projected/04810067-c268-4cae-905c-bb4913ae908e-kube-api-access-zx7tk\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.433811 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-ring-data-devices\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.433845 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-swiftconf\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.433914 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-scripts\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.434228 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/04810067-c268-4cae-905c-bb4913ae908e-etc-swift\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.434731 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-scripts\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.434971 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-ring-data-devices\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.438708 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-dispersionconf\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.439905 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-swiftconf\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.452384 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7tk\" (UniqueName: \"kubernetes.io/projected/04810067-c268-4cae-905c-bb4913ae908e-kube-api-access-zx7tk\") pod \"swift-ring-rebalance-debug-np59w\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.538370 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:58:56 crc kubenswrapper[4692]: I0309 09:58:56.960538 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-np59w"] Mar 09 09:58:57 crc kubenswrapper[4692]: I0309 09:58:57.826731 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" event={"ID":"04810067-c268-4cae-905c-bb4913ae908e","Type":"ContainerStarted","Data":"49c59ddc3778b11939e3d214d4ca8519b433ac2550c237dcc16a09d5e064ff7a"} Mar 09 09:58:57 crc kubenswrapper[4692]: I0309 09:58:57.827076 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" event={"ID":"04810067-c268-4cae-905c-bb4913ae908e","Type":"ContainerStarted","Data":"c6d19d3a714daa11c67d62fd1f7d17779fefe43d57f7b41d86199bca9ee5b57c"} Mar 09 09:58:57 crc kubenswrapper[4692]: I0309 09:58:57.843382 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" podStartSLOduration=1.843367107 podStartE2EDuration="1.843367107s" podCreationTimestamp="2026-03-09 09:58:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:58:57.839812016 +0000 UTC m=+2338.664547597" watchObservedRunningTime="2026-03-09 09:58:57.843367107 +0000 UTC m=+2338.668102698" Mar 09 09:58:58 crc kubenswrapper[4692]: I0309 09:58:58.071734 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:58:58 crc kubenswrapper[4692]: E0309 09:58:58.071953 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:58:58 crc kubenswrapper[4692]: I0309 09:58:58.838835 4692 generic.go:334] "Generic (PLEG): container finished" podID="04810067-c268-4cae-905c-bb4913ae908e" containerID="49c59ddc3778b11939e3d214d4ca8519b433ac2550c237dcc16a09d5e064ff7a" exitCode=0 Mar 09 09:58:58 crc kubenswrapper[4692]: I0309 09:58:58.838910 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" event={"ID":"04810067-c268-4cae-905c-bb4913ae908e","Type":"ContainerDied","Data":"49c59ddc3778b11939e3d214d4ca8519b433ac2550c237dcc16a09d5e064ff7a"} Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.178208 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.209505 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-np59w"] Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.223144 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-np59w"] Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.300628 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-ring-data-devices\") pod \"04810067-c268-4cae-905c-bb4913ae908e\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.300780 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-scripts\") pod \"04810067-c268-4cae-905c-bb4913ae908e\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.300849 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-swiftconf\") pod \"04810067-c268-4cae-905c-bb4913ae908e\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.300875 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/04810067-c268-4cae-905c-bb4913ae908e-etc-swift\") pod \"04810067-c268-4cae-905c-bb4913ae908e\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.300904 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-dispersionconf\") pod \"04810067-c268-4cae-905c-bb4913ae908e\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.300959 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx7tk\" (UniqueName: \"kubernetes.io/projected/04810067-c268-4cae-905c-bb4913ae908e-kube-api-access-zx7tk\") pod \"04810067-c268-4cae-905c-bb4913ae908e\" (UID: \"04810067-c268-4cae-905c-bb4913ae908e\") " Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.301573 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "04810067-c268-4cae-905c-bb4913ae908e" (UID: "04810067-c268-4cae-905c-bb4913ae908e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.302084 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04810067-c268-4cae-905c-bb4913ae908e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "04810067-c268-4cae-905c-bb4913ae908e" (UID: "04810067-c268-4cae-905c-bb4913ae908e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.307492 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04810067-c268-4cae-905c-bb4913ae908e-kube-api-access-zx7tk" (OuterVolumeSpecName: "kube-api-access-zx7tk") pod "04810067-c268-4cae-905c-bb4913ae908e" (UID: "04810067-c268-4cae-905c-bb4913ae908e"). InnerVolumeSpecName "kube-api-access-zx7tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.327802 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "04810067-c268-4cae-905c-bb4913ae908e" (UID: "04810067-c268-4cae-905c-bb4913ae908e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.328049 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "04810067-c268-4cae-905c-bb4913ae908e" (UID: "04810067-c268-4cae-905c-bb4913ae908e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.333280 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-scripts" (OuterVolumeSpecName: "scripts") pod "04810067-c268-4cae-905c-bb4913ae908e" (UID: "04810067-c268-4cae-905c-bb4913ae908e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.402455 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.402489 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.402498 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/04810067-c268-4cae-905c-bb4913ae908e-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.402508 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/04810067-c268-4cae-905c-bb4913ae908e-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.402520 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx7tk\" (UniqueName: \"kubernetes.io/projected/04810067-c268-4cae-905c-bb4913ae908e-kube-api-access-zx7tk\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.402529 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/04810067-c268-4cae-905c-bb4913ae908e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.872997 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6d19d3a714daa11c67d62fd1f7d17779fefe43d57f7b41d86199bca9ee5b57c" Mar 09 09:59:00 crc kubenswrapper[4692]: I0309 09:59:00.873105 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-np59w" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.369125 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-72b4s"] Mar 09 09:59:01 crc kubenswrapper[4692]: E0309 09:59:01.369498 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04810067-c268-4cae-905c-bb4913ae908e" containerName="swift-ring-rebalance" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.369516 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="04810067-c268-4cae-905c-bb4913ae908e" containerName="swift-ring-rebalance" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.369705 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="04810067-c268-4cae-905c-bb4913ae908e" containerName="swift-ring-rebalance" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.370416 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.374389 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.379575 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.387638 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-72b4s"] Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.417801 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfpk\" (UniqueName: \"kubernetes.io/projected/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-kube-api-access-sdfpk\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.417852 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-etc-swift\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.417890 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-scripts\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.418083 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-swiftconf\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.418194 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-dispersionconf\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.418242 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-ring-data-devices\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.520408 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfpk\" (UniqueName: \"kubernetes.io/projected/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-kube-api-access-sdfpk\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.520488 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-etc-swift\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.520541 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-scripts\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.520588 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-swiftconf\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.520623 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-dispersionconf\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.520648 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-ring-data-devices\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.521374 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-etc-swift\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.521961 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-ring-data-devices\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.522087 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-scripts\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.525823 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-dispersionconf\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.525823 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-swiftconf\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.550642 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfpk\" (UniqueName: \"kubernetes.io/projected/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-kube-api-access-sdfpk\") pod \"swift-ring-rebalance-debug-72b4s\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.690157 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:01 crc kubenswrapper[4692]: I0309 09:59:01.923065 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-72b4s"] Mar 09 09:59:02 crc kubenswrapper[4692]: I0309 09:59:02.097768 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04810067-c268-4cae-905c-bb4913ae908e" path="/var/lib/kubelet/pods/04810067-c268-4cae-905c-bb4913ae908e/volumes" Mar 09 09:59:02 crc kubenswrapper[4692]: I0309 09:59:02.896461 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" event={"ID":"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331","Type":"ContainerStarted","Data":"cd7921286faa7a351505044018ae816ea4516eafda0caaaf0762e9442f8f5bf6"} Mar 09 09:59:02 crc kubenswrapper[4692]: I0309 09:59:02.896842 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" event={"ID":"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331","Type":"ContainerStarted","Data":"a6179d24397b31ee29f7cafc8ad5d5b2752cc8a3eadbce0303d59466190c2a75"} Mar 09 09:59:02 crc kubenswrapper[4692]: I0309 09:59:02.922697 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" podStartSLOduration=1.9226730029999999 podStartE2EDuration="1.922673003s" podCreationTimestamp="2026-03-09 09:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:02.920307046 +0000 UTC m=+2343.745042647" watchObservedRunningTime="2026-03-09 09:59:02.922673003 +0000 UTC m=+2343.747408574" Mar 09 09:59:03 crc kubenswrapper[4692]: I0309 09:59:03.910049 4692 generic.go:334] "Generic (PLEG): container finished" podID="8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" containerID="cd7921286faa7a351505044018ae816ea4516eafda0caaaf0762e9442f8f5bf6" exitCode=0 Mar 09 09:59:03 crc kubenswrapper[4692]: I0309 09:59:03.910118 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" event={"ID":"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331","Type":"ContainerDied","Data":"cd7921286faa7a351505044018ae816ea4516eafda0caaaf0762e9442f8f5bf6"} Mar 09 09:59:04 crc kubenswrapper[4692]: I0309 09:59:04.786199 4692 scope.go:117] "RemoveContainer" containerID="3580d1770362f5765d81516c9e8b50f135dd442db13b96320b675908b45da6f5" Mar 09 09:59:04 crc kubenswrapper[4692]: I0309 09:59:04.830395 4692 scope.go:117] "RemoveContainer" containerID="4517cad5bcb4f554212542d663de526a9492034948b19012e9a59f0451a37bf9" Mar 09 09:59:04 crc kubenswrapper[4692]: I0309 09:59:04.868097 4692 scope.go:117] "RemoveContainer" containerID="a7480c36d0baf51cfa88845f2aec3a3e54a7632d4c9e48c893bfbf48500cd0af" Mar 09 09:59:04 crc kubenswrapper[4692]: I0309 09:59:04.906524 4692 scope.go:117] "RemoveContainer" containerID="96195b47de9285eeebb5eafe3541bf95f85f1b855a554db9c83ee01c5bd7b5ea" Mar 09 09:59:04 crc kubenswrapper[4692]: I0309 09:59:04.954619 4692 scope.go:117] "RemoveContainer" containerID="09b01c5676c4a787f0f0da4a2d13ab72cd517be381ded7a3b8734bd995dfe863" Mar 09 09:59:04 crc kubenswrapper[4692]: I0309 09:59:04.989922 4692 scope.go:117] "RemoveContainer" containerID="13a00efb8b3e99458e89f1a1e1ecbaca90137b55c6f8660cb667647a57cf6f2d" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.041614 4692 scope.go:117] "RemoveContainer" containerID="209c9fe69490d2c566df58b2617fe4e6708f47d94bf73b96abf6de28064b84dc" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.152472 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.180853 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-72b4s"] Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.188047 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-72b4s"] Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.293469 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-ring-data-devices\") pod \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.293536 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-scripts\") pod \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.293599 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-etc-swift\") pod \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.293751 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdfpk\" (UniqueName: \"kubernetes.io/projected/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-kube-api-access-sdfpk\") pod \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.293777 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-swiftconf\") pod \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.293800 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-dispersionconf\") pod \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\" (UID: \"8f3ebe0b-eca8-46f8-a22c-2ba2abaae331\") " Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.294096 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" (UID: "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.294216 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.295423 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" (UID: "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.300645 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-kube-api-access-sdfpk" (OuterVolumeSpecName: "kube-api-access-sdfpk") pod "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" (UID: "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331"). InnerVolumeSpecName "kube-api-access-sdfpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.319230 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-scripts" (OuterVolumeSpecName: "scripts") pod "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" (UID: "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.322143 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" (UID: "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.323668 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" (UID: "8f3ebe0b-eca8-46f8-a22c-2ba2abaae331"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.395215 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdfpk\" (UniqueName: \"kubernetes.io/projected/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-kube-api-access-sdfpk\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.395249 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.395258 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.395267 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.395277 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.937236 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6179d24397b31ee29f7cafc8ad5d5b2752cc8a3eadbce0303d59466190c2a75" Mar 09 09:59:05 crc kubenswrapper[4692]: I0309 09:59:05.937285 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-72b4s" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.081054 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" path="/var/lib/kubelet/pods/8f3ebe0b-eca8-46f8-a22c-2ba2abaae331/volumes" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.324948 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb"] Mar 09 09:59:06 crc kubenswrapper[4692]: E0309 09:59:06.325269 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" containerName="swift-ring-rebalance" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.325283 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" containerName="swift-ring-rebalance" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.325494 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f3ebe0b-eca8-46f8-a22c-2ba2abaae331" containerName="swift-ring-rebalance" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.325981 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.327997 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.328514 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.339780 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb"] Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.410567 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-ring-data-devices\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.410628 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-dispersionconf\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.410675 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdcde1cb-9705-4c55-95ba-b532f669f724-etc-swift\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.410708 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-scripts\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.410903 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2vvt\" (UniqueName: \"kubernetes.io/projected/bdcde1cb-9705-4c55-95ba-b532f669f724-kube-api-access-j2vvt\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.411185 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-swiftconf\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.513137 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-ring-data-devices\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.513227 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-dispersionconf\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.513260 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdcde1cb-9705-4c55-95ba-b532f669f724-etc-swift\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.513288 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-scripts\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.513316 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2vvt\" (UniqueName: \"kubernetes.io/projected/bdcde1cb-9705-4c55-95ba-b532f669f724-kube-api-access-j2vvt\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.513349 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-swiftconf\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.514045 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdcde1cb-9705-4c55-95ba-b532f669f724-etc-swift\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.514259 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-ring-data-devices\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.516088 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-scripts\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.519328 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-dispersionconf\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.527324 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-swiftconf\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.536525 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2vvt\" (UniqueName: \"kubernetes.io/projected/bdcde1cb-9705-4c55-95ba-b532f669f724-kube-api-access-j2vvt\") pod \"swift-ring-rebalance-debug-wr6mb\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:06 crc kubenswrapper[4692]: I0309 09:59:06.642551 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:07 crc kubenswrapper[4692]: I0309 09:59:07.087389 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb"] Mar 09 09:59:07 crc kubenswrapper[4692]: I0309 09:59:07.956913 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" event={"ID":"bdcde1cb-9705-4c55-95ba-b532f669f724","Type":"ContainerStarted","Data":"c39a6ec6b5787e28d96f0698d28fc4e4749cef8d8f7381f41cb412e99181770e"} Mar 09 09:59:07 crc kubenswrapper[4692]: I0309 09:59:07.957935 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" event={"ID":"bdcde1cb-9705-4c55-95ba-b532f669f724","Type":"ContainerStarted","Data":"45207503150a5ff5f1c2fba6a885ff254117a704fbc07ec57a5d37a6fa67e823"} Mar 09 09:59:07 crc kubenswrapper[4692]: I0309 09:59:07.974073 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" podStartSLOduration=1.974044958 podStartE2EDuration="1.974044958s" podCreationTimestamp="2026-03-09 09:59:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:07.97236196 +0000 UTC m=+2348.797097541" watchObservedRunningTime="2026-03-09 09:59:07.974044958 +0000 UTC m=+2348.798780539" Mar 09 09:59:08 crc kubenswrapper[4692]: I0309 09:59:08.966296 4692 generic.go:334] "Generic (PLEG): container finished" podID="bdcde1cb-9705-4c55-95ba-b532f669f724" containerID="c39a6ec6b5787e28d96f0698d28fc4e4749cef8d8f7381f41cb412e99181770e" exitCode=0 Mar 09 09:59:08 crc kubenswrapper[4692]: I0309 09:59:08.966333 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" event={"ID":"bdcde1cb-9705-4c55-95ba-b532f669f724","Type":"ContainerDied","Data":"c39a6ec6b5787e28d96f0698d28fc4e4749cef8d8f7381f41cb412e99181770e"} Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.251786 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.297834 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb"] Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.308910 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb"] Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.382849 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2vvt\" (UniqueName: \"kubernetes.io/projected/bdcde1cb-9705-4c55-95ba-b532f669f724-kube-api-access-j2vvt\") pod \"bdcde1cb-9705-4c55-95ba-b532f669f724\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.382983 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdcde1cb-9705-4c55-95ba-b532f669f724-etc-swift\") pod \"bdcde1cb-9705-4c55-95ba-b532f669f724\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.383040 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-ring-data-devices\") pod \"bdcde1cb-9705-4c55-95ba-b532f669f724\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.383115 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-scripts\") pod \"bdcde1cb-9705-4c55-95ba-b532f669f724\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.383221 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-swiftconf\") pod \"bdcde1cb-9705-4c55-95ba-b532f669f724\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.383249 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-dispersionconf\") pod \"bdcde1cb-9705-4c55-95ba-b532f669f724\" (UID: \"bdcde1cb-9705-4c55-95ba-b532f669f724\") " Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.392885 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bdcde1cb-9705-4c55-95ba-b532f669f724" (UID: "bdcde1cb-9705-4c55-95ba-b532f669f724"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.392969 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdcde1cb-9705-4c55-95ba-b532f669f724-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bdcde1cb-9705-4c55-95ba-b532f669f724" (UID: "bdcde1cb-9705-4c55-95ba-b532f669f724"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.403044 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdcde1cb-9705-4c55-95ba-b532f669f724-kube-api-access-j2vvt" (OuterVolumeSpecName: "kube-api-access-j2vvt") pod "bdcde1cb-9705-4c55-95ba-b532f669f724" (UID: "bdcde1cb-9705-4c55-95ba-b532f669f724"). InnerVolumeSpecName "kube-api-access-j2vvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.432756 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-scripts" (OuterVolumeSpecName: "scripts") pod "bdcde1cb-9705-4c55-95ba-b532f669f724" (UID: "bdcde1cb-9705-4c55-95ba-b532f669f724"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.433403 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bdcde1cb-9705-4c55-95ba-b532f669f724" (UID: "bdcde1cb-9705-4c55-95ba-b532f669f724"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.449827 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bdcde1cb-9705-4c55-95ba-b532f669f724" (UID: "bdcde1cb-9705-4c55-95ba-b532f669f724"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.485071 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdcde1cb-9705-4c55-95ba-b532f669f724-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.485123 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.485138 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdcde1cb-9705-4c55-95ba-b532f669f724-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.485150 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.485180 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdcde1cb-9705-4c55-95ba-b532f669f724-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.485194 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2vvt\" (UniqueName: \"kubernetes.io/projected/bdcde1cb-9705-4c55-95ba-b532f669f724-kube-api-access-j2vvt\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.983229 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45207503150a5ff5f1c2fba6a885ff254117a704fbc07ec57a5d37a6fa67e823" Mar 09 09:59:10 crc kubenswrapper[4692]: I0309 09:59:10.983312 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wr6mb" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.072336 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:59:11 crc kubenswrapper[4692]: E0309 09:59:11.072571 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.470018 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz"] Mar 09 09:59:11 crc kubenswrapper[4692]: E0309 09:59:11.472838 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdcde1cb-9705-4c55-95ba-b532f669f724" containerName="swift-ring-rebalance" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.473086 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdcde1cb-9705-4c55-95ba-b532f669f724" containerName="swift-ring-rebalance" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.473356 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdcde1cb-9705-4c55-95ba-b532f669f724" containerName="swift-ring-rebalance" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.474097 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.479224 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz"] Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.479656 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.479941 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.600649 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-swiftconf\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.600787 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-ring-data-devices\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.600875 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-scripts\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.600908 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/687f192a-55ac-4b37-a65c-9e8b4994323f-etc-swift\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.601535 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-dispersionconf\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.601602 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvfpj\" (UniqueName: \"kubernetes.io/projected/687f192a-55ac-4b37-a65c-9e8b4994323f-kube-api-access-jvfpj\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.703147 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvfpj\" (UniqueName: \"kubernetes.io/projected/687f192a-55ac-4b37-a65c-9e8b4994323f-kube-api-access-jvfpj\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.703242 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-swiftconf\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.703270 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-ring-data-devices\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.703338 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-scripts\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.703365 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/687f192a-55ac-4b37-a65c-9e8b4994323f-etc-swift\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.703390 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-dispersionconf\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.704288 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/687f192a-55ac-4b37-a65c-9e8b4994323f-etc-swift\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.704686 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-ring-data-devices\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.705507 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-scripts\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.708888 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-dispersionconf\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.712896 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-swiftconf\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.721518 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvfpj\" (UniqueName: \"kubernetes.io/projected/687f192a-55ac-4b37-a65c-9e8b4994323f-kube-api-access-jvfpj\") pod \"swift-ring-rebalance-debug-wbwgz\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:11 crc kubenswrapper[4692]: I0309 09:59:11.801546 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:12 crc kubenswrapper[4692]: I0309 09:59:12.037395 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz"] Mar 09 09:59:12 crc kubenswrapper[4692]: I0309 09:59:12.081401 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdcde1cb-9705-4c55-95ba-b532f669f724" path="/var/lib/kubelet/pods/bdcde1cb-9705-4c55-95ba-b532f669f724/volumes" Mar 09 09:59:13 crc kubenswrapper[4692]: I0309 09:59:13.003771 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" event={"ID":"687f192a-55ac-4b37-a65c-9e8b4994323f","Type":"ContainerStarted","Data":"28a0514e56ea08d97f779af95ccabbff6fccf035fa321903c56314dbceb835a3"} Mar 09 09:59:13 crc kubenswrapper[4692]: I0309 09:59:13.005730 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" event={"ID":"687f192a-55ac-4b37-a65c-9e8b4994323f","Type":"ContainerStarted","Data":"50b84c1c53d60dd8aa54a53489cffa6a6de1323d455b53922334109641c26a47"} Mar 09 09:59:13 crc kubenswrapper[4692]: I0309 09:59:13.027738 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" podStartSLOduration=2.027718059 podStartE2EDuration="2.027718059s" podCreationTimestamp="2026-03-09 09:59:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:13.023000356 +0000 UTC m=+2353.847735947" watchObservedRunningTime="2026-03-09 09:59:13.027718059 +0000 UTC m=+2353.852453640" Mar 09 09:59:14 crc kubenswrapper[4692]: I0309 09:59:14.015504 4692 generic.go:334] "Generic (PLEG): container finished" podID="687f192a-55ac-4b37-a65c-9e8b4994323f" containerID="28a0514e56ea08d97f779af95ccabbff6fccf035fa321903c56314dbceb835a3" exitCode=0 Mar 09 09:59:14 crc kubenswrapper[4692]: I0309 09:59:14.015566 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" event={"ID":"687f192a-55ac-4b37-a65c-9e8b4994323f","Type":"ContainerDied","Data":"28a0514e56ea08d97f779af95ccabbff6fccf035fa321903c56314dbceb835a3"} Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.372858 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.408287 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz"] Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.413998 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz"] Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.460932 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-scripts\") pod \"687f192a-55ac-4b37-a65c-9e8b4994323f\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.460979 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-swiftconf\") pod \"687f192a-55ac-4b37-a65c-9e8b4994323f\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.461029 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/687f192a-55ac-4b37-a65c-9e8b4994323f-etc-swift\") pod \"687f192a-55ac-4b37-a65c-9e8b4994323f\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.461127 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvfpj\" (UniqueName: \"kubernetes.io/projected/687f192a-55ac-4b37-a65c-9e8b4994323f-kube-api-access-jvfpj\") pod \"687f192a-55ac-4b37-a65c-9e8b4994323f\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.461184 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-dispersionconf\") pod \"687f192a-55ac-4b37-a65c-9e8b4994323f\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.461208 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-ring-data-devices\") pod \"687f192a-55ac-4b37-a65c-9e8b4994323f\" (UID: \"687f192a-55ac-4b37-a65c-9e8b4994323f\") " Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.461921 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "687f192a-55ac-4b37-a65c-9e8b4994323f" (UID: "687f192a-55ac-4b37-a65c-9e8b4994323f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.462219 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/687f192a-55ac-4b37-a65c-9e8b4994323f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "687f192a-55ac-4b37-a65c-9e8b4994323f" (UID: "687f192a-55ac-4b37-a65c-9e8b4994323f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.466375 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/687f192a-55ac-4b37-a65c-9e8b4994323f-kube-api-access-jvfpj" (OuterVolumeSpecName: "kube-api-access-jvfpj") pod "687f192a-55ac-4b37-a65c-9e8b4994323f" (UID: "687f192a-55ac-4b37-a65c-9e8b4994323f"). InnerVolumeSpecName "kube-api-access-jvfpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.481315 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-scripts" (OuterVolumeSpecName: "scripts") pod "687f192a-55ac-4b37-a65c-9e8b4994323f" (UID: "687f192a-55ac-4b37-a65c-9e8b4994323f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.482886 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "687f192a-55ac-4b37-a65c-9e8b4994323f" (UID: "687f192a-55ac-4b37-a65c-9e8b4994323f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.483857 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "687f192a-55ac-4b37-a65c-9e8b4994323f" (UID: "687f192a-55ac-4b37-a65c-9e8b4994323f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.564306 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.564348 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.564360 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/687f192a-55ac-4b37-a65c-9e8b4994323f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.564373 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvfpj\" (UniqueName: \"kubernetes.io/projected/687f192a-55ac-4b37-a65c-9e8b4994323f-kube-api-access-jvfpj\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.564388 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/687f192a-55ac-4b37-a65c-9e8b4994323f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:15 crc kubenswrapper[4692]: I0309 09:59:15.564399 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/687f192a-55ac-4b37-a65c-9e8b4994323f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.034709 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50b84c1c53d60dd8aa54a53489cffa6a6de1323d455b53922334109641c26a47" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.034764 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wbwgz" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.081443 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="687f192a-55ac-4b37-a65c-9e8b4994323f" path="/var/lib/kubelet/pods/687f192a-55ac-4b37-a65c-9e8b4994323f/volumes" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.553453 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2"] Mar 09 09:59:16 crc kubenswrapper[4692]: E0309 09:59:16.553782 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687f192a-55ac-4b37-a65c-9e8b4994323f" containerName="swift-ring-rebalance" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.553795 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="687f192a-55ac-4b37-a65c-9e8b4994323f" containerName="swift-ring-rebalance" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.553931 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="687f192a-55ac-4b37-a65c-9e8b4994323f" containerName="swift-ring-rebalance" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.554533 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.556471 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.557077 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.573173 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2"] Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.678761 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpjl\" (UniqueName: \"kubernetes.io/projected/6fe5624b-261d-4690-8979-25cff1e7910a-kube-api-access-xmpjl\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.678816 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-dispersionconf\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.679034 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-scripts\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.679149 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6fe5624b-261d-4690-8979-25cff1e7910a-etc-swift\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.679197 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-ring-data-devices\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.679294 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-swiftconf\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.780880 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6fe5624b-261d-4690-8979-25cff1e7910a-etc-swift\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.780962 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-ring-data-devices\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.781024 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-swiftconf\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.781105 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpjl\" (UniqueName: \"kubernetes.io/projected/6fe5624b-261d-4690-8979-25cff1e7910a-kube-api-access-xmpjl\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.781153 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-dispersionconf\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.781260 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-scripts\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.781642 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6fe5624b-261d-4690-8979-25cff1e7910a-etc-swift\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.782036 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-ring-data-devices\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.782330 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-scripts\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.785949 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-dispersionconf\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.794015 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-swiftconf\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.796667 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpjl\" (UniqueName: \"kubernetes.io/projected/6fe5624b-261d-4690-8979-25cff1e7910a-kube-api-access-xmpjl\") pod \"swift-ring-rebalance-debug-7vqc2\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:16 crc kubenswrapper[4692]: I0309 09:59:16.873862 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:17 crc kubenswrapper[4692]: I0309 09:59:17.284260 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2"] Mar 09 09:59:18 crc kubenswrapper[4692]: I0309 09:59:18.055175 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" event={"ID":"6fe5624b-261d-4690-8979-25cff1e7910a","Type":"ContainerStarted","Data":"302c063ddd7a1cd0bbb5da39cec9cc0d61214b8f1872ffad288902a0341b4d1c"} Mar 09 09:59:18 crc kubenswrapper[4692]: I0309 09:59:18.055612 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" event={"ID":"6fe5624b-261d-4690-8979-25cff1e7910a","Type":"ContainerStarted","Data":"a15f0c2a9f0e14dbdffd6674f7f8b90adff8d9abff9d3b3ef1c5c07cc1617743"} Mar 09 09:59:18 crc kubenswrapper[4692]: I0309 09:59:18.082001 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" podStartSLOduration=2.081977396 podStartE2EDuration="2.081977396s" podCreationTimestamp="2026-03-09 09:59:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:18.078003354 +0000 UTC m=+2358.902739035" watchObservedRunningTime="2026-03-09 09:59:18.081977396 +0000 UTC m=+2358.906712977" Mar 09 09:59:19 crc kubenswrapper[4692]: I0309 09:59:19.073051 4692 generic.go:334] "Generic (PLEG): container finished" podID="6fe5624b-261d-4690-8979-25cff1e7910a" containerID="302c063ddd7a1cd0bbb5da39cec9cc0d61214b8f1872ffad288902a0341b4d1c" exitCode=0 Mar 09 09:59:19 crc kubenswrapper[4692]: I0309 09:59:19.073120 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" event={"ID":"6fe5624b-261d-4690-8979-25cff1e7910a","Type":"ContainerDied","Data":"302c063ddd7a1cd0bbb5da39cec9cc0d61214b8f1872ffad288902a0341b4d1c"} Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.389277 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.434190 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2"] Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.442151 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2"] Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.446355 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmpjl\" (UniqueName: \"kubernetes.io/projected/6fe5624b-261d-4690-8979-25cff1e7910a-kube-api-access-xmpjl\") pod \"6fe5624b-261d-4690-8979-25cff1e7910a\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.446433 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-dispersionconf\") pod \"6fe5624b-261d-4690-8979-25cff1e7910a\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.446497 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6fe5624b-261d-4690-8979-25cff1e7910a-etc-swift\") pod \"6fe5624b-261d-4690-8979-25cff1e7910a\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.446606 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-swiftconf\") pod \"6fe5624b-261d-4690-8979-25cff1e7910a\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.446645 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-ring-data-devices\") pod \"6fe5624b-261d-4690-8979-25cff1e7910a\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.446684 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-scripts\") pod \"6fe5624b-261d-4690-8979-25cff1e7910a\" (UID: \"6fe5624b-261d-4690-8979-25cff1e7910a\") " Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.448048 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fe5624b-261d-4690-8979-25cff1e7910a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6fe5624b-261d-4690-8979-25cff1e7910a" (UID: "6fe5624b-261d-4690-8979-25cff1e7910a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.449740 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6fe5624b-261d-4690-8979-25cff1e7910a" (UID: "6fe5624b-261d-4690-8979-25cff1e7910a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.460422 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fe5624b-261d-4690-8979-25cff1e7910a-kube-api-access-xmpjl" (OuterVolumeSpecName: "kube-api-access-xmpjl") pod "6fe5624b-261d-4690-8979-25cff1e7910a" (UID: "6fe5624b-261d-4690-8979-25cff1e7910a"). InnerVolumeSpecName "kube-api-access-xmpjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.473285 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-scripts" (OuterVolumeSpecName: "scripts") pod "6fe5624b-261d-4690-8979-25cff1e7910a" (UID: "6fe5624b-261d-4690-8979-25cff1e7910a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.485963 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6fe5624b-261d-4690-8979-25cff1e7910a" (UID: "6fe5624b-261d-4690-8979-25cff1e7910a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.486630 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6fe5624b-261d-4690-8979-25cff1e7910a" (UID: "6fe5624b-261d-4690-8979-25cff1e7910a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.548911 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6fe5624b-261d-4690-8979-25cff1e7910a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.549152 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.549237 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.549325 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6fe5624b-261d-4690-8979-25cff1e7910a-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.549382 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmpjl\" (UniqueName: \"kubernetes.io/projected/6fe5624b-261d-4690-8979-25cff1e7910a-kube-api-access-xmpjl\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:20 crc kubenswrapper[4692]: I0309 09:59:20.549437 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6fe5624b-261d-4690-8979-25cff1e7910a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.094376 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a15f0c2a9f0e14dbdffd6674f7f8b90adff8d9abff9d3b3ef1c5c07cc1617743" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.094413 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7vqc2" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.591080 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xln7p"] Mar 09 09:59:21 crc kubenswrapper[4692]: E0309 09:59:21.591543 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe5624b-261d-4690-8979-25cff1e7910a" containerName="swift-ring-rebalance" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.591560 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe5624b-261d-4690-8979-25cff1e7910a" containerName="swift-ring-rebalance" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.591731 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe5624b-261d-4690-8979-25cff1e7910a" containerName="swift-ring-rebalance" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.592362 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.595034 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.596439 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.612901 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xln7p"] Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.667714 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4cjn\" (UniqueName: \"kubernetes.io/projected/d1e6ef35-7c41-4abc-9849-387c714a3f1d-kube-api-access-q4cjn\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.667795 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-dispersionconf\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.667866 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d1e6ef35-7c41-4abc-9849-387c714a3f1d-etc-swift\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.667910 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.667936 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-swiftconf\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.668179 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-scripts\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.770519 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d1e6ef35-7c41-4abc-9849-387c714a3f1d-etc-swift\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.770621 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.770650 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-swiftconf\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.770691 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-scripts\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.770762 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4cjn\" (UniqueName: \"kubernetes.io/projected/d1e6ef35-7c41-4abc-9849-387c714a3f1d-kube-api-access-q4cjn\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.770810 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-dispersionconf\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.771132 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d1e6ef35-7c41-4abc-9849-387c714a3f1d-etc-swift\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.771701 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-ring-data-devices\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.771834 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-scripts\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.775957 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-swiftconf\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.776331 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-dispersionconf\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.821149 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4cjn\" (UniqueName: \"kubernetes.io/projected/d1e6ef35-7c41-4abc-9849-387c714a3f1d-kube-api-access-q4cjn\") pod \"swift-ring-rebalance-debug-xln7p\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:21 crc kubenswrapper[4692]: I0309 09:59:21.918110 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:22 crc kubenswrapper[4692]: I0309 09:59:22.087124 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fe5624b-261d-4690-8979-25cff1e7910a" path="/var/lib/kubelet/pods/6fe5624b-261d-4690-8979-25cff1e7910a/volumes" Mar 09 09:59:22 crc kubenswrapper[4692]: I0309 09:59:22.166022 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xln7p"] Mar 09 09:59:22 crc kubenswrapper[4692]: W0309 09:59:22.173524 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1e6ef35_7c41_4abc_9849_387c714a3f1d.slice/crio-e37f963206bd569b4b85614cc7e5deca25e3644ac7721aed11238e5e2b72dc81 WatchSource:0}: Error finding container e37f963206bd569b4b85614cc7e5deca25e3644ac7721aed11238e5e2b72dc81: Status 404 returned error can't find the container with id e37f963206bd569b4b85614cc7e5deca25e3644ac7721aed11238e5e2b72dc81 Mar 09 09:59:23 crc kubenswrapper[4692]: I0309 09:59:23.126210 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" event={"ID":"d1e6ef35-7c41-4abc-9849-387c714a3f1d","Type":"ContainerStarted","Data":"07effac9ba4e8781ed2de446c7438adbb4e43cfacf7d414a2b66da716ac5f33c"} Mar 09 09:59:23 crc kubenswrapper[4692]: I0309 09:59:23.126254 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" event={"ID":"d1e6ef35-7c41-4abc-9849-387c714a3f1d","Type":"ContainerStarted","Data":"e37f963206bd569b4b85614cc7e5deca25e3644ac7721aed11238e5e2b72dc81"} Mar 09 09:59:23 crc kubenswrapper[4692]: I0309 09:59:23.151593 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" podStartSLOduration=2.151571788 podStartE2EDuration="2.151571788s" podCreationTimestamp="2026-03-09 09:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:23.144114587 +0000 UTC m=+2363.968850178" watchObservedRunningTime="2026-03-09 09:59:23.151571788 +0000 UTC m=+2363.976307369" Mar 09 09:59:24 crc kubenswrapper[4692]: I0309 09:59:24.136542 4692 generic.go:334] "Generic (PLEG): container finished" podID="d1e6ef35-7c41-4abc-9849-387c714a3f1d" containerID="07effac9ba4e8781ed2de446c7438adbb4e43cfacf7d414a2b66da716ac5f33c" exitCode=0 Mar 09 09:59:24 crc kubenswrapper[4692]: I0309 09:59:24.137609 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" event={"ID":"d1e6ef35-7c41-4abc-9849-387c714a3f1d","Type":"ContainerDied","Data":"07effac9ba4e8781ed2de446c7438adbb4e43cfacf7d414a2b66da716ac5f33c"} Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.418871 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.467871 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xln7p"] Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.475731 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-xln7p"] Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.527848 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4cjn\" (UniqueName: \"kubernetes.io/projected/d1e6ef35-7c41-4abc-9849-387c714a3f1d-kube-api-access-q4cjn\") pod \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.527922 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-ring-data-devices\") pod \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.528034 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-scripts\") pod \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.528082 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-dispersionconf\") pod \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.528843 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "d1e6ef35-7c41-4abc-9849-387c714a3f1d" (UID: "d1e6ef35-7c41-4abc-9849-387c714a3f1d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.528933 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d1e6ef35-7c41-4abc-9849-387c714a3f1d-etc-swift\") pod \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.529653 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1e6ef35-7c41-4abc-9849-387c714a3f1d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d1e6ef35-7c41-4abc-9849-387c714a3f1d" (UID: "d1e6ef35-7c41-4abc-9849-387c714a3f1d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.529733 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-swiftconf\") pod \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\" (UID: \"d1e6ef35-7c41-4abc-9849-387c714a3f1d\") " Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.530380 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/d1e6ef35-7c41-4abc-9849-387c714a3f1d-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.530411 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.536434 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1e6ef35-7c41-4abc-9849-387c714a3f1d-kube-api-access-q4cjn" (OuterVolumeSpecName: "kube-api-access-q4cjn") pod "d1e6ef35-7c41-4abc-9849-387c714a3f1d" (UID: "d1e6ef35-7c41-4abc-9849-387c714a3f1d"). InnerVolumeSpecName "kube-api-access-q4cjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.552947 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-scripts" (OuterVolumeSpecName: "scripts") pod "d1e6ef35-7c41-4abc-9849-387c714a3f1d" (UID: "d1e6ef35-7c41-4abc-9849-387c714a3f1d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.558767 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "d1e6ef35-7c41-4abc-9849-387c714a3f1d" (UID: "d1e6ef35-7c41-4abc-9849-387c714a3f1d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.562769 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "d1e6ef35-7c41-4abc-9849-387c714a3f1d" (UID: "d1e6ef35-7c41-4abc-9849-387c714a3f1d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.631782 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4cjn\" (UniqueName: \"kubernetes.io/projected/d1e6ef35-7c41-4abc-9849-387c714a3f1d-kube-api-access-q4cjn\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.631830 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1e6ef35-7c41-4abc-9849-387c714a3f1d-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.631846 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:25 crc kubenswrapper[4692]: I0309 09:59:25.631855 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/d1e6ef35-7c41-4abc-9849-387c714a3f1d-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.071772 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:59:26 crc kubenswrapper[4692]: E0309 09:59:26.072077 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.084325 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1e6ef35-7c41-4abc-9849-387c714a3f1d" path="/var/lib/kubelet/pods/d1e6ef35-7c41-4abc-9849-387c714a3f1d/volumes" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.164020 4692 scope.go:117] "RemoveContainer" containerID="07effac9ba4e8781ed2de446c7438adbb4e43cfacf7d414a2b66da716ac5f33c" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.164109 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-xln7p" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.610455 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s"] Mar 09 09:59:26 crc kubenswrapper[4692]: E0309 09:59:26.611279 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1e6ef35-7c41-4abc-9849-387c714a3f1d" containerName="swift-ring-rebalance" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.611295 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1e6ef35-7c41-4abc-9849-387c714a3f1d" containerName="swift-ring-rebalance" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.611467 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1e6ef35-7c41-4abc-9849-387c714a3f1d" containerName="swift-ring-rebalance" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.612289 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.614232 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.614502 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.622598 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s"] Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.646075 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-dispersionconf\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.646189 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44t9d\" (UniqueName: \"kubernetes.io/projected/ad916c26-a778-4939-878c-4afb49c97b71-kube-api-access-44t9d\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.646304 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-swiftconf\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.646357 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad916c26-a778-4939-878c-4afb49c97b71-etc-swift\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.646459 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-ring-data-devices\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.646510 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-scripts\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.748701 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-ring-data-devices\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.748789 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-scripts\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.748879 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-dispersionconf\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.748959 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44t9d\" (UniqueName: \"kubernetes.io/projected/ad916c26-a778-4939-878c-4afb49c97b71-kube-api-access-44t9d\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.749002 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-swiftconf\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.749023 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad916c26-a778-4939-878c-4afb49c97b71-etc-swift\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.749868 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad916c26-a778-4939-878c-4afb49c97b71-etc-swift\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.750000 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-scripts\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.750045 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-ring-data-devices\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.753609 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-dispersionconf\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.757203 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-swiftconf\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.772111 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44t9d\" (UniqueName: \"kubernetes.io/projected/ad916c26-a778-4939-878c-4afb49c97b71-kube-api-access-44t9d\") pod \"swift-ring-rebalance-debug-wdx5s\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:26 crc kubenswrapper[4692]: I0309 09:59:26.936712 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:27 crc kubenswrapper[4692]: I0309 09:59:27.152522 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s"] Mar 09 09:59:27 crc kubenswrapper[4692]: I0309 09:59:27.179524 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" event={"ID":"ad916c26-a778-4939-878c-4afb49c97b71","Type":"ContainerStarted","Data":"d5f7820d4f6c54ee6e2b920fc3c7161c78f1df8c711887b06fc847173ea52633"} Mar 09 09:59:28 crc kubenswrapper[4692]: I0309 09:59:28.195134 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" event={"ID":"ad916c26-a778-4939-878c-4afb49c97b71","Type":"ContainerStarted","Data":"b3e83c2f9748e85094f461c07e8e14fe2f5a9c383f06cab0ff4f24a491577c46"} Mar 09 09:59:28 crc kubenswrapper[4692]: I0309 09:59:28.235837 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" podStartSLOduration=2.235811043 podStartE2EDuration="2.235811043s" podCreationTimestamp="2026-03-09 09:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:28.2215462 +0000 UTC m=+2369.046281851" watchObservedRunningTime="2026-03-09 09:59:28.235811043 +0000 UTC m=+2369.060546624" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.213714 4692 generic.go:334] "Generic (PLEG): container finished" podID="ad916c26-a778-4939-878c-4afb49c97b71" containerID="b3e83c2f9748e85094f461c07e8e14fe2f5a9c383f06cab0ff4f24a491577c46" exitCode=0 Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.213825 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" event={"ID":"ad916c26-a778-4939-878c-4afb49c97b71","Type":"ContainerDied","Data":"b3e83c2f9748e85094f461c07e8e14fe2f5a9c383f06cab0ff4f24a491577c46"} Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.454517 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qnh6g"] Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.456458 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.466075 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnh6g"] Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.499240 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqqh8\" (UniqueName: \"kubernetes.io/projected/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-kube-api-access-mqqh8\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.499328 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-utilities\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.499394 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-catalog-content\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.601398 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-catalog-content\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.601518 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqqh8\" (UniqueName: \"kubernetes.io/projected/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-kube-api-access-mqqh8\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.601595 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-utilities\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.602258 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-utilities\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.602566 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-catalog-content\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.635808 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqqh8\" (UniqueName: \"kubernetes.io/projected/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-kube-api-access-mqqh8\") pod \"redhat-operators-qnh6g\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:29 crc kubenswrapper[4692]: I0309 09:59:29.787648 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.275778 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qnh6g"] Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.511177 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.554789 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s"] Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.561896 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s"] Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.617405 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-swiftconf\") pod \"ad916c26-a778-4939-878c-4afb49c97b71\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.617624 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-scripts\") pod \"ad916c26-a778-4939-878c-4afb49c97b71\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.617779 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad916c26-a778-4939-878c-4afb49c97b71-etc-swift\") pod \"ad916c26-a778-4939-878c-4afb49c97b71\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.617898 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44t9d\" (UniqueName: \"kubernetes.io/projected/ad916c26-a778-4939-878c-4afb49c97b71-kube-api-access-44t9d\") pod \"ad916c26-a778-4939-878c-4afb49c97b71\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.617941 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-dispersionconf\") pod \"ad916c26-a778-4939-878c-4afb49c97b71\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.617974 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-ring-data-devices\") pod \"ad916c26-a778-4939-878c-4afb49c97b71\" (UID: \"ad916c26-a778-4939-878c-4afb49c97b71\") " Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.618906 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ad916c26-a778-4939-878c-4afb49c97b71" (UID: "ad916c26-a778-4939-878c-4afb49c97b71"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.619188 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad916c26-a778-4939-878c-4afb49c97b71-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ad916c26-a778-4939-878c-4afb49c97b71" (UID: "ad916c26-a778-4939-878c-4afb49c97b71"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.640925 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad916c26-a778-4939-878c-4afb49c97b71-kube-api-access-44t9d" (OuterVolumeSpecName: "kube-api-access-44t9d") pod "ad916c26-a778-4939-878c-4afb49c97b71" (UID: "ad916c26-a778-4939-878c-4afb49c97b71"). InnerVolumeSpecName "kube-api-access-44t9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.644619 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-scripts" (OuterVolumeSpecName: "scripts") pod "ad916c26-a778-4939-878c-4afb49c97b71" (UID: "ad916c26-a778-4939-878c-4afb49c97b71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.652763 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ad916c26-a778-4939-878c-4afb49c97b71" (UID: "ad916c26-a778-4939-878c-4afb49c97b71"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.673261 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ad916c26-a778-4939-878c-4afb49c97b71" (UID: "ad916c26-a778-4939-878c-4afb49c97b71"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.720145 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ad916c26-a778-4939-878c-4afb49c97b71-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.720211 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44t9d\" (UniqueName: \"kubernetes.io/projected/ad916c26-a778-4939-878c-4afb49c97b71-kube-api-access-44t9d\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.720227 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.720239 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.720286 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ad916c26-a778-4939-878c-4afb49c97b71-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:30 crc kubenswrapper[4692]: I0309 09:59:30.720300 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ad916c26-a778-4939-878c-4afb49c97b71-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.245682 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wdx5s" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.245655 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5f7820d4f6c54ee6e2b920fc3c7161c78f1df8c711887b06fc847173ea52633" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.248029 4692 generic.go:334] "Generic (PLEG): container finished" podID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerID="9e4c47eb60160a072989b197bb5e5cd82f76897aa43865b259519002efa5d9ee" exitCode=0 Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.248208 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnh6g" event={"ID":"92f1158a-ddbf-4fd5-b446-c24f01cb3c66","Type":"ContainerDied","Data":"9e4c47eb60160a072989b197bb5e5cd82f76897aa43865b259519002efa5d9ee"} Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.248311 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnh6g" event={"ID":"92f1158a-ddbf-4fd5-b446-c24f01cb3c66","Type":"ContainerStarted","Data":"bd54af5aaf1cc5243078d61350a033eb4b88eae63b4809ba3ba314d048c18713"} Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.691519 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd"] Mar 09 09:59:31 crc kubenswrapper[4692]: E0309 09:59:31.692284 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad916c26-a778-4939-878c-4afb49c97b71" containerName="swift-ring-rebalance" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.692302 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad916c26-a778-4939-878c-4afb49c97b71" containerName="swift-ring-rebalance" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.692495 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad916c26-a778-4939-878c-4afb49c97b71" containerName="swift-ring-rebalance" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.693115 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.694789 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.695261 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.707314 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd"] Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.736984 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-scripts\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.737051 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-ring-data-devices\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.737077 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstnv\" (UniqueName: \"kubernetes.io/projected/65005dbc-a595-49bc-a53e-477f38b5bcaa-kube-api-access-zstnv\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.737097 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-swiftconf\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.737179 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-dispersionconf\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.737243 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65005dbc-a595-49bc-a53e-477f38b5bcaa-etc-swift\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.838365 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-ring-data-devices\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.838415 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstnv\" (UniqueName: \"kubernetes.io/projected/65005dbc-a595-49bc-a53e-477f38b5bcaa-kube-api-access-zstnv\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.838441 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-swiftconf\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.838477 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-dispersionconf\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.838523 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65005dbc-a595-49bc-a53e-477f38b5bcaa-etc-swift\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.838577 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-scripts\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.839071 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-ring-data-devices\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.839353 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-scripts\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.839383 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65005dbc-a595-49bc-a53e-477f38b5bcaa-etc-swift\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.844043 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-dispersionconf\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.852059 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-swiftconf\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:31 crc kubenswrapper[4692]: I0309 09:59:31.854767 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstnv\" (UniqueName: \"kubernetes.io/projected/65005dbc-a595-49bc-a53e-477f38b5bcaa-kube-api-access-zstnv\") pod \"swift-ring-rebalance-debug-7kzsd\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:32 crc kubenswrapper[4692]: I0309 09:59:32.043135 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:32 crc kubenswrapper[4692]: I0309 09:59:32.081001 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad916c26-a778-4939-878c-4afb49c97b71" path="/var/lib/kubelet/pods/ad916c26-a778-4939-878c-4afb49c97b71/volumes" Mar 09 09:59:32 crc kubenswrapper[4692]: I0309 09:59:32.447010 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd"] Mar 09 09:59:32 crc kubenswrapper[4692]: W0309 09:59:32.468355 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65005dbc_a595_49bc_a53e_477f38b5bcaa.slice/crio-4a6e0ab1030b85a121a3d927503c5896b1bffd5020d73c882d3ddda55cb7b440 WatchSource:0}: Error finding container 4a6e0ab1030b85a121a3d927503c5896b1bffd5020d73c882d3ddda55cb7b440: Status 404 returned error can't find the container with id 4a6e0ab1030b85a121a3d927503c5896b1bffd5020d73c882d3ddda55cb7b440 Mar 09 09:59:33 crc kubenswrapper[4692]: I0309 09:59:33.274824 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnh6g" event={"ID":"92f1158a-ddbf-4fd5-b446-c24f01cb3c66","Type":"ContainerStarted","Data":"126c24a5be08051918fe7fecca21526d4ae585125d39ec6f3a11383c47efb92f"} Mar 09 09:59:33 crc kubenswrapper[4692]: I0309 09:59:33.277635 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" event={"ID":"65005dbc-a595-49bc-a53e-477f38b5bcaa","Type":"ContainerStarted","Data":"f8e56ebc5937b1471152d1fad98b9aaecfbd7686e2869329cdc8b4341a23dbca"} Mar 09 09:59:33 crc kubenswrapper[4692]: I0309 09:59:33.277662 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" event={"ID":"65005dbc-a595-49bc-a53e-477f38b5bcaa","Type":"ContainerStarted","Data":"4a6e0ab1030b85a121a3d927503c5896b1bffd5020d73c882d3ddda55cb7b440"} Mar 09 09:59:33 crc kubenswrapper[4692]: I0309 09:59:33.324678 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" podStartSLOduration=2.324658119 podStartE2EDuration="2.324658119s" podCreationTimestamp="2026-03-09 09:59:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:33.319023859 +0000 UTC m=+2374.143759450" watchObservedRunningTime="2026-03-09 09:59:33.324658119 +0000 UTC m=+2374.149393700" Mar 09 09:59:34 crc kubenswrapper[4692]: I0309 09:59:34.286951 4692 generic.go:334] "Generic (PLEG): container finished" podID="65005dbc-a595-49bc-a53e-477f38b5bcaa" containerID="f8e56ebc5937b1471152d1fad98b9aaecfbd7686e2869329cdc8b4341a23dbca" exitCode=0 Mar 09 09:59:34 crc kubenswrapper[4692]: I0309 09:59:34.287029 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" event={"ID":"65005dbc-a595-49bc-a53e-477f38b5bcaa","Type":"ContainerDied","Data":"f8e56ebc5937b1471152d1fad98b9aaecfbd7686e2869329cdc8b4341a23dbca"} Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.296746 4692 generic.go:334] "Generic (PLEG): container finished" podID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerID="126c24a5be08051918fe7fecca21526d4ae585125d39ec6f3a11383c47efb92f" exitCode=0 Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.296807 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnh6g" event={"ID":"92f1158a-ddbf-4fd5-b446-c24f01cb3c66","Type":"ContainerDied","Data":"126c24a5be08051918fe7fecca21526d4ae585125d39ec6f3a11383c47efb92f"} Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.598074 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.635094 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd"] Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.644074 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd"] Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.700471 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zstnv\" (UniqueName: \"kubernetes.io/projected/65005dbc-a595-49bc-a53e-477f38b5bcaa-kube-api-access-zstnv\") pod \"65005dbc-a595-49bc-a53e-477f38b5bcaa\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.700553 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-scripts\") pod \"65005dbc-a595-49bc-a53e-477f38b5bcaa\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.700587 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-ring-data-devices\") pod \"65005dbc-a595-49bc-a53e-477f38b5bcaa\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.700608 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-swiftconf\") pod \"65005dbc-a595-49bc-a53e-477f38b5bcaa\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.700640 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65005dbc-a595-49bc-a53e-477f38b5bcaa-etc-swift\") pod \"65005dbc-a595-49bc-a53e-477f38b5bcaa\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.700684 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-dispersionconf\") pod \"65005dbc-a595-49bc-a53e-477f38b5bcaa\" (UID: \"65005dbc-a595-49bc-a53e-477f38b5bcaa\") " Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.701318 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "65005dbc-a595-49bc-a53e-477f38b5bcaa" (UID: "65005dbc-a595-49bc-a53e-477f38b5bcaa"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.701499 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65005dbc-a595-49bc-a53e-477f38b5bcaa-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "65005dbc-a595-49bc-a53e-477f38b5bcaa" (UID: "65005dbc-a595-49bc-a53e-477f38b5bcaa"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.706241 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65005dbc-a595-49bc-a53e-477f38b5bcaa-kube-api-access-zstnv" (OuterVolumeSpecName: "kube-api-access-zstnv") pod "65005dbc-a595-49bc-a53e-477f38b5bcaa" (UID: "65005dbc-a595-49bc-a53e-477f38b5bcaa"). InnerVolumeSpecName "kube-api-access-zstnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.722665 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-scripts" (OuterVolumeSpecName: "scripts") pod "65005dbc-a595-49bc-a53e-477f38b5bcaa" (UID: "65005dbc-a595-49bc-a53e-477f38b5bcaa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.725099 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "65005dbc-a595-49bc-a53e-477f38b5bcaa" (UID: "65005dbc-a595-49bc-a53e-477f38b5bcaa"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.733332 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "65005dbc-a595-49bc-a53e-477f38b5bcaa" (UID: "65005dbc-a595-49bc-a53e-477f38b5bcaa"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.802845 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zstnv\" (UniqueName: \"kubernetes.io/projected/65005dbc-a595-49bc-a53e-477f38b5bcaa-kube-api-access-zstnv\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.802895 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.802914 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65005dbc-a595-49bc-a53e-477f38b5bcaa-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.802926 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.802938 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65005dbc-a595-49bc-a53e-477f38b5bcaa-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:35 crc kubenswrapper[4692]: I0309 09:59:35.802951 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65005dbc-a595-49bc-a53e-477f38b5bcaa-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.081987 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65005dbc-a595-49bc-a53e-477f38b5bcaa" path="/var/lib/kubelet/pods/65005dbc-a595-49bc-a53e-477f38b5bcaa/volumes" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.308362 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnh6g" event={"ID":"92f1158a-ddbf-4fd5-b446-c24f01cb3c66","Type":"ContainerStarted","Data":"bcc9d2be9e7b3f78819bca2d860979de2cf66a8b3630e8b524dc3af9b7a9eff7"} Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.311535 4692 scope.go:117] "RemoveContainer" containerID="f8e56ebc5937b1471152d1fad98b9aaecfbd7686e2869329cdc8b4341a23dbca" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.311567 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7kzsd" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.331129 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qnh6g" podStartSLOduration=2.5505372299999998 podStartE2EDuration="7.331104881s" podCreationTimestamp="2026-03-09 09:59:29 +0000 UTC" firstStartedPulling="2026-03-09 09:59:31.251002263 +0000 UTC m=+2372.075737834" lastFinishedPulling="2026-03-09 09:59:36.031569904 +0000 UTC m=+2376.856305485" observedRunningTime="2026-03-09 09:59:36.326905092 +0000 UTC m=+2377.151640683" watchObservedRunningTime="2026-03-09 09:59:36.331104881 +0000 UTC m=+2377.155840462" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.807292 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622n6"] Mar 09 09:59:36 crc kubenswrapper[4692]: E0309 09:59:36.807695 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65005dbc-a595-49bc-a53e-477f38b5bcaa" containerName="swift-ring-rebalance" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.807718 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="65005dbc-a595-49bc-a53e-477f38b5bcaa" containerName="swift-ring-rebalance" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.807932 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="65005dbc-a595-49bc-a53e-477f38b5bcaa" containerName="swift-ring-rebalance" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.808593 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.811220 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.812419 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.826732 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622n6"] Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.919296 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-swiftconf\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.919342 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpczr\" (UniqueName: \"kubernetes.io/projected/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-kube-api-access-qpczr\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.919388 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-dispersionconf\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.919406 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-ring-data-devices\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.919471 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-scripts\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:36 crc kubenswrapper[4692]: I0309 09:59:36.919507 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-etc-swift\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.021232 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-ring-data-devices\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.021312 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-scripts\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.021384 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-etc-swift\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.021469 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-swiftconf\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.021495 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpczr\" (UniqueName: \"kubernetes.io/projected/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-kube-api-access-qpczr\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.021564 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-dispersionconf\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.022485 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-etc-swift\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.023348 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-ring-data-devices\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.023969 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-scripts\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.028193 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-dispersionconf\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.028193 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-swiftconf\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.044324 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpczr\" (UniqueName: \"kubernetes.io/projected/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-kube-api-access-qpczr\") pod \"swift-ring-rebalance-debug-622n6\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.127563 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:37 crc kubenswrapper[4692]: I0309 09:59:37.580876 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622n6"] Mar 09 09:59:37 crc kubenswrapper[4692]: W0309 09:59:37.593816 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf8378bf_ed9b_46cc_abce_ccfda6dddbe2.slice/crio-9cf4ba31df3d95bfc18ff1cf0a3589ef1cf9d55d201734d3635ad621a214ff96 WatchSource:0}: Error finding container 9cf4ba31df3d95bfc18ff1cf0a3589ef1cf9d55d201734d3635ad621a214ff96: Status 404 returned error can't find the container with id 9cf4ba31df3d95bfc18ff1cf0a3589ef1cf9d55d201734d3635ad621a214ff96 Mar 09 09:59:38 crc kubenswrapper[4692]: I0309 09:59:38.340635 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" event={"ID":"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2","Type":"ContainerStarted","Data":"4a4c70e4f836ea6b0175628e30c293bfc7961aae14c4694f0f0a62375499e543"} Mar 09 09:59:38 crc kubenswrapper[4692]: I0309 09:59:38.340994 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" event={"ID":"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2","Type":"ContainerStarted","Data":"9cf4ba31df3d95bfc18ff1cf0a3589ef1cf9d55d201734d3635ad621a214ff96"} Mar 09 09:59:38 crc kubenswrapper[4692]: I0309 09:59:38.365600 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" podStartSLOduration=2.36557132 podStartE2EDuration="2.36557132s" podCreationTimestamp="2026-03-09 09:59:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:38.358743647 +0000 UTC m=+2379.183479268" watchObservedRunningTime="2026-03-09 09:59:38.36557132 +0000 UTC m=+2379.190306921" Mar 09 09:59:39 crc kubenswrapper[4692]: I0309 09:59:39.071425 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:59:39 crc kubenswrapper[4692]: E0309 09:59:39.071842 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:59:39 crc kubenswrapper[4692]: I0309 09:59:39.789585 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:39 crc kubenswrapper[4692]: I0309 09:59:39.790794 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:40 crc kubenswrapper[4692]: I0309 09:59:40.366684 4692 generic.go:334] "Generic (PLEG): container finished" podID="cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" containerID="4a4c70e4f836ea6b0175628e30c293bfc7961aae14c4694f0f0a62375499e543" exitCode=0 Mar 09 09:59:40 crc kubenswrapper[4692]: I0309 09:59:40.366774 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" event={"ID":"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2","Type":"ContainerDied","Data":"4a4c70e4f836ea6b0175628e30c293bfc7961aae14c4694f0f0a62375499e543"} Mar 09 09:59:40 crc kubenswrapper[4692]: I0309 09:59:40.846035 4692 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qnh6g" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="registry-server" probeResult="failure" output=< Mar 09 09:59:40 crc kubenswrapper[4692]: timeout: failed to connect service ":50051" within 1s Mar 09 09:59:40 crc kubenswrapper[4692]: > Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.700254 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.752356 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpczr\" (UniqueName: \"kubernetes.io/projected/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-kube-api-access-qpczr\") pod \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.752428 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-scripts\") pod \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.752463 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-ring-data-devices\") pod \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.752504 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-dispersionconf\") pod \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.752542 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-etc-swift\") pod \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.752589 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-swiftconf\") pod \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\" (UID: \"cf8378bf-ed9b-46cc-abce-ccfda6dddbe2\") " Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.754924 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" (UID: "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.755125 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" (UID: "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.774683 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622n6"] Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.784405 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-kube-api-access-qpczr" (OuterVolumeSpecName: "kube-api-access-qpczr") pod "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" (UID: "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2"). InnerVolumeSpecName "kube-api-access-qpczr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.787214 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" (UID: "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.789260 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-scripts" (OuterVolumeSpecName: "scripts") pod "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" (UID: "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.790873 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" (UID: "cf8378bf-ed9b-46cc-abce-ccfda6dddbe2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.794504 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-622n6"] Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.854342 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.854397 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.854410 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.854427 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.854442 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:41 crc kubenswrapper[4692]: I0309 09:59:41.854452 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpczr\" (UniqueName: \"kubernetes.io/projected/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2-kube-api-access-qpczr\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.080943 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" path="/var/lib/kubelet/pods/cf8378bf-ed9b-46cc-abce-ccfda6dddbe2/volumes" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.388665 4692 scope.go:117] "RemoveContainer" containerID="4a4c70e4f836ea6b0175628e30c293bfc7961aae14c4694f0f0a62375499e543" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.388703 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-622n6" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.905656 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-62smz"] Mar 09 09:59:42 crc kubenswrapper[4692]: E0309 09:59:42.906125 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" containerName="swift-ring-rebalance" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.906144 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" containerName="swift-ring-rebalance" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.906383 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf8378bf-ed9b-46cc-abce-ccfda6dddbe2" containerName="swift-ring-rebalance" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.907140 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.909112 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.910646 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.922430 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-62smz"] Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.975069 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-dispersionconf\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.975207 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-ring-data-devices\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.975228 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-etc-swift\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.975253 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-scripts\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.975280 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-swiftconf\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:42 crc kubenswrapper[4692]: I0309 09:59:42.975347 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7xrs\" (UniqueName: \"kubernetes.io/projected/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-kube-api-access-n7xrs\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.075821 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-dispersionconf\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.075914 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-ring-data-devices\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.075936 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-etc-swift\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.075970 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-scripts\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.076000 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-swiftconf\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.076019 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7xrs\" (UniqueName: \"kubernetes.io/projected/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-kube-api-access-n7xrs\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.077010 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-etc-swift\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.077303 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-ring-data-devices\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.077328 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-scripts\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.081071 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-swiftconf\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.083586 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-dispersionconf\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.094741 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7xrs\" (UniqueName: \"kubernetes.io/projected/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-kube-api-access-n7xrs\") pod \"swift-ring-rebalance-debug-62smz\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.226007 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:43 crc kubenswrapper[4692]: I0309 09:59:43.576378 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-62smz"] Mar 09 09:59:44 crc kubenswrapper[4692]: I0309 09:59:44.419272 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" event={"ID":"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd","Type":"ContainerStarted","Data":"b6bf8462ad09bac39f0c708f34d236c6cb983e2b531ebe205c4055b651c8609d"} Mar 09 09:59:44 crc kubenswrapper[4692]: I0309 09:59:44.419434 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" event={"ID":"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd","Type":"ContainerStarted","Data":"4c2c84e266e02e20065f8f50add3f6c14a04e893f80390af19b50c0b6fe3f9cf"} Mar 09 09:59:44 crc kubenswrapper[4692]: I0309 09:59:44.468020 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" podStartSLOduration=2.467994785 podStartE2EDuration="2.467994785s" podCreationTimestamp="2026-03-09 09:59:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:44.466606676 +0000 UTC m=+2385.291342257" watchObservedRunningTime="2026-03-09 09:59:44.467994785 +0000 UTC m=+2385.292730366" Mar 09 09:59:45 crc kubenswrapper[4692]: I0309 09:59:45.432535 4692 generic.go:334] "Generic (PLEG): container finished" podID="a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" containerID="b6bf8462ad09bac39f0c708f34d236c6cb983e2b531ebe205c4055b651c8609d" exitCode=0 Mar 09 09:59:45 crc kubenswrapper[4692]: I0309 09:59:45.432624 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" event={"ID":"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd","Type":"ContainerDied","Data":"b6bf8462ad09bac39f0c708f34d236c6cb983e2b531ebe205c4055b651c8609d"} Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.768607 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.818313 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-62smz"] Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.824738 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-62smz"] Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.879858 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-dispersionconf\") pod \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.880020 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-swiftconf\") pod \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.880109 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7xrs\" (UniqueName: \"kubernetes.io/projected/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-kube-api-access-n7xrs\") pod \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.880144 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-ring-data-devices\") pod \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.880183 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-scripts\") pod \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.880285 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-etc-swift\") pod \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\" (UID: \"a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd\") " Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.881292 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" (UID: "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.882311 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" (UID: "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.883118 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.883186 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.890886 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-kube-api-access-n7xrs" (OuterVolumeSpecName: "kube-api-access-n7xrs") pod "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" (UID: "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd"). InnerVolumeSpecName "kube-api-access-n7xrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.914473 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" (UID: "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.917475 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-scripts" (OuterVolumeSpecName: "scripts") pod "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" (UID: "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.917536 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" (UID: "a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.985471 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.985520 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.985530 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:46 crc kubenswrapper[4692]: I0309 09:59:46.985542 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7xrs\" (UniqueName: \"kubernetes.io/projected/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd-kube-api-access-n7xrs\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.458924 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c2c84e266e02e20065f8f50add3f6c14a04e893f80390af19b50c0b6fe3f9cf" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.459002 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-62smz" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.953285 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj"] Mar 09 09:59:47 crc kubenswrapper[4692]: E0309 09:59:47.953696 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" containerName="swift-ring-rebalance" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.953712 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" containerName="swift-ring-rebalance" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.953850 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" containerName="swift-ring-rebalance" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.954369 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.959653 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.960380 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:47 crc kubenswrapper[4692]: I0309 09:59:47.971587 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj"] Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.002290 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-ring-data-devices\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.002376 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwsr5\" (UniqueName: \"kubernetes.io/projected/7047f39c-a10c-434c-8a1b-e76597767d04-kube-api-access-zwsr5\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.002437 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7047f39c-a10c-434c-8a1b-e76597767d04-etc-swift\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.002469 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-dispersionconf\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.002505 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-scripts\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.002561 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-swiftconf\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.082863 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd" path="/var/lib/kubelet/pods/a5026a19-fcde-4e80-bbdc-2dbfbdfe8ffd/volumes" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.103595 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-swiftconf\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.103719 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-ring-data-devices\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.103761 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwsr5\" (UniqueName: \"kubernetes.io/projected/7047f39c-a10c-434c-8a1b-e76597767d04-kube-api-access-zwsr5\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.103840 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7047f39c-a10c-434c-8a1b-e76597767d04-etc-swift\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.103868 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-dispersionconf\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.103908 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-scripts\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.105748 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7047f39c-a10c-434c-8a1b-e76597767d04-etc-swift\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.105819 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-scripts\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.106191 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-ring-data-devices\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.112771 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-dispersionconf\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.112756 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-swiftconf\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.135830 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwsr5\" (UniqueName: \"kubernetes.io/projected/7047f39c-a10c-434c-8a1b-e76597767d04-kube-api-access-zwsr5\") pod \"swift-ring-rebalance-debug-vpmfj\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.275298 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:48 crc kubenswrapper[4692]: I0309 09:59:48.737152 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj"] Mar 09 09:59:49 crc kubenswrapper[4692]: I0309 09:59:49.484737 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" event={"ID":"7047f39c-a10c-434c-8a1b-e76597767d04","Type":"ContainerStarted","Data":"f1306d0da5e4830d8de1b4fc81607f914910060502a6adb159fb417fe3c3ee66"} Mar 09 09:59:49 crc kubenswrapper[4692]: I0309 09:59:49.485501 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" event={"ID":"7047f39c-a10c-434c-8a1b-e76597767d04","Type":"ContainerStarted","Data":"a66a9beca8c5141bc67e3e673704b7b8b430a9ce8e526976590de79eb41e131f"} Mar 09 09:59:49 crc kubenswrapper[4692]: I0309 09:59:49.842426 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:49 crc kubenswrapper[4692]: I0309 09:59:49.867603 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" podStartSLOduration=2.867579403 podStartE2EDuration="2.867579403s" podCreationTimestamp="2026-03-09 09:59:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:49.51187219 +0000 UTC m=+2390.336607771" watchObservedRunningTime="2026-03-09 09:59:49.867579403 +0000 UTC m=+2390.692314974" Mar 09 09:59:49 crc kubenswrapper[4692]: I0309 09:59:49.900288 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:50 crc kubenswrapper[4692]: I0309 09:59:50.498464 4692 generic.go:334] "Generic (PLEG): container finished" podID="7047f39c-a10c-434c-8a1b-e76597767d04" containerID="f1306d0da5e4830d8de1b4fc81607f914910060502a6adb159fb417fe3c3ee66" exitCode=0 Mar 09 09:59:50 crc kubenswrapper[4692]: I0309 09:59:50.498539 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" event={"ID":"7047f39c-a10c-434c-8a1b-e76597767d04","Type":"ContainerDied","Data":"f1306d0da5e4830d8de1b4fc81607f914910060502a6adb159fb417fe3c3ee66"} Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.828900 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.868753 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj"] Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.878609 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj"] Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.971390 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-ring-data-devices\") pod \"7047f39c-a10c-434c-8a1b-e76597767d04\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.971452 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-swiftconf\") pod \"7047f39c-a10c-434c-8a1b-e76597767d04\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.971471 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-scripts\") pod \"7047f39c-a10c-434c-8a1b-e76597767d04\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.971516 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwsr5\" (UniqueName: \"kubernetes.io/projected/7047f39c-a10c-434c-8a1b-e76597767d04-kube-api-access-zwsr5\") pod \"7047f39c-a10c-434c-8a1b-e76597767d04\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.971613 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-dispersionconf\") pod \"7047f39c-a10c-434c-8a1b-e76597767d04\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.972677 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "7047f39c-a10c-434c-8a1b-e76597767d04" (UID: "7047f39c-a10c-434c-8a1b-e76597767d04"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.972739 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7047f39c-a10c-434c-8a1b-e76597767d04-etc-swift\") pod \"7047f39c-a10c-434c-8a1b-e76597767d04\" (UID: \"7047f39c-a10c-434c-8a1b-e76597767d04\") " Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.973548 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7047f39c-a10c-434c-8a1b-e76597767d04-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "7047f39c-a10c-434c-8a1b-e76597767d04" (UID: "7047f39c-a10c-434c-8a1b-e76597767d04"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.974015 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.982422 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7047f39c-a10c-434c-8a1b-e76597767d04-kube-api-access-zwsr5" (OuterVolumeSpecName: "kube-api-access-zwsr5") pod "7047f39c-a10c-434c-8a1b-e76597767d04" (UID: "7047f39c-a10c-434c-8a1b-e76597767d04"). InnerVolumeSpecName "kube-api-access-zwsr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.996491 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-scripts" (OuterVolumeSpecName: "scripts") pod "7047f39c-a10c-434c-8a1b-e76597767d04" (UID: "7047f39c-a10c-434c-8a1b-e76597767d04"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:51 crc kubenswrapper[4692]: I0309 09:59:51.997832 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "7047f39c-a10c-434c-8a1b-e76597767d04" (UID: "7047f39c-a10c-434c-8a1b-e76597767d04"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.001707 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "7047f39c-a10c-434c-8a1b-e76597767d04" (UID: "7047f39c-a10c-434c-8a1b-e76597767d04"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.072288 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 09:59:52 crc kubenswrapper[4692]: E0309 09:59:52.072618 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.075367 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/7047f39c-a10c-434c-8a1b-e76597767d04-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.075412 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.075423 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7047f39c-a10c-434c-8a1b-e76597767d04-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.075437 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwsr5\" (UniqueName: \"kubernetes.io/projected/7047f39c-a10c-434c-8a1b-e76597767d04-kube-api-access-zwsr5\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.075451 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/7047f39c-a10c-434c-8a1b-e76597767d04-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.083952 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7047f39c-a10c-434c-8a1b-e76597767d04" path="/var/lib/kubelet/pods/7047f39c-a10c-434c-8a1b-e76597767d04/volumes" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.522757 4692 scope.go:117] "RemoveContainer" containerID="f1306d0da5e4830d8de1b4fc81607f914910060502a6adb159fb417fe3c3ee66" Mar 09 09:59:52 crc kubenswrapper[4692]: I0309 09:59:52.522902 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vpmfj" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.021828 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx"] Mar 09 09:59:53 crc kubenswrapper[4692]: E0309 09:59:53.022695 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7047f39c-a10c-434c-8a1b-e76597767d04" containerName="swift-ring-rebalance" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.022710 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="7047f39c-a10c-434c-8a1b-e76597767d04" containerName="swift-ring-rebalance" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.024768 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="7047f39c-a10c-434c-8a1b-e76597767d04" containerName="swift-ring-rebalance" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.025535 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.028056 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.030746 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.038905 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx"] Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.194616 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jknwj\" (UniqueName: \"kubernetes.io/projected/bdc6dd32-00a7-4164-8f35-320700382b1f-kube-api-access-jknwj\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.194702 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-swiftconf\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.194752 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdc6dd32-00a7-4164-8f35-320700382b1f-etc-swift\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.194776 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-dispersionconf\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.194921 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-scripts\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.194995 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-ring-data-devices\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.296913 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jknwj\" (UniqueName: \"kubernetes.io/projected/bdc6dd32-00a7-4164-8f35-320700382b1f-kube-api-access-jknwj\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.297000 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-swiftconf\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.297037 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdc6dd32-00a7-4164-8f35-320700382b1f-etc-swift\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.297058 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-dispersionconf\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.297130 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-scripts\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.297210 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-ring-data-devices\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.297940 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdc6dd32-00a7-4164-8f35-320700382b1f-etc-swift\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.298710 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-scripts\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.299299 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-ring-data-devices\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.303238 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-swiftconf\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.303507 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-dispersionconf\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.317819 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jknwj\" (UniqueName: \"kubernetes.io/projected/bdc6dd32-00a7-4164-8f35-320700382b1f-kube-api-access-jknwj\") pod \"swift-ring-rebalance-debug-qrxcx\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.363347 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qnh6g"] Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.363646 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qnh6g" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="registry-server" containerID="cri-o://bcc9d2be9e7b3f78819bca2d860979de2cf66a8b3630e8b524dc3af9b7a9eff7" gracePeriod=2 Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.366630 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.539102 4692 generic.go:334] "Generic (PLEG): container finished" podID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerID="bcc9d2be9e7b3f78819bca2d860979de2cf66a8b3630e8b524dc3af9b7a9eff7" exitCode=0 Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.539185 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnh6g" event={"ID":"92f1158a-ddbf-4fd5-b446-c24f01cb3c66","Type":"ContainerDied","Data":"bcc9d2be9e7b3f78819bca2d860979de2cf66a8b3630e8b524dc3af9b7a9eff7"} Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.791361 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.886945 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx"] Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.904833 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-catalog-content\") pod \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.904965 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-utilities\") pod \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.905023 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqqh8\" (UniqueName: \"kubernetes.io/projected/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-kube-api-access-mqqh8\") pod \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\" (UID: \"92f1158a-ddbf-4fd5-b446-c24f01cb3c66\") " Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.906060 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-utilities" (OuterVolumeSpecName: "utilities") pod "92f1158a-ddbf-4fd5-b446-c24f01cb3c66" (UID: "92f1158a-ddbf-4fd5-b446-c24f01cb3c66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:53 crc kubenswrapper[4692]: I0309 09:59:53.913517 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-kube-api-access-mqqh8" (OuterVolumeSpecName: "kube-api-access-mqqh8") pod "92f1158a-ddbf-4fd5-b446-c24f01cb3c66" (UID: "92f1158a-ddbf-4fd5-b446-c24f01cb3c66"). InnerVolumeSpecName "kube-api-access-mqqh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.006562 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.007022 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqqh8\" (UniqueName: \"kubernetes.io/projected/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-kube-api-access-mqqh8\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.056437 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92f1158a-ddbf-4fd5-b446-c24f01cb3c66" (UID: "92f1158a-ddbf-4fd5-b446-c24f01cb3c66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.108375 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92f1158a-ddbf-4fd5-b446-c24f01cb3c66-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.553459 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" event={"ID":"bdc6dd32-00a7-4164-8f35-320700382b1f","Type":"ContainerStarted","Data":"8b57adaff31d14e0799698f3f2242fbd193f96cca51f751415e62c82941bfd4c"} Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.553844 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" event={"ID":"bdc6dd32-00a7-4164-8f35-320700382b1f","Type":"ContainerStarted","Data":"1d7b8581d07bd3ca6e1b1f21fc2842f76774c9535322b232feeb6af828051f3c"} Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.557417 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qnh6g" event={"ID":"92f1158a-ddbf-4fd5-b446-c24f01cb3c66","Type":"ContainerDied","Data":"bd54af5aaf1cc5243078d61350a033eb4b88eae63b4809ba3ba314d048c18713"} Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.557487 4692 scope.go:117] "RemoveContainer" containerID="bcc9d2be9e7b3f78819bca2d860979de2cf66a8b3630e8b524dc3af9b7a9eff7" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.557930 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qnh6g" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.578028 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" podStartSLOduration=1.578011043 podStartE2EDuration="1.578011043s" podCreationTimestamp="2026-03-09 09:59:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:54.575137021 +0000 UTC m=+2395.399872632" watchObservedRunningTime="2026-03-09 09:59:54.578011043 +0000 UTC m=+2395.402746624" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.605468 4692 scope.go:117] "RemoveContainer" containerID="126c24a5be08051918fe7fecca21526d4ae585125d39ec6f3a11383c47efb92f" Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.615260 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qnh6g"] Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.626205 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qnh6g"] Mar 09 09:59:54 crc kubenswrapper[4692]: I0309 09:59:54.634719 4692 scope.go:117] "RemoveContainer" containerID="9e4c47eb60160a072989b197bb5e5cd82f76897aa43865b259519002efa5d9ee" Mar 09 09:59:55 crc kubenswrapper[4692]: I0309 09:59:55.574143 4692 generic.go:334] "Generic (PLEG): container finished" podID="bdc6dd32-00a7-4164-8f35-320700382b1f" containerID="8b57adaff31d14e0799698f3f2242fbd193f96cca51f751415e62c82941bfd4c" exitCode=0 Mar 09 09:59:55 crc kubenswrapper[4692]: I0309 09:59:55.574237 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" event={"ID":"bdc6dd32-00a7-4164-8f35-320700382b1f","Type":"ContainerDied","Data":"8b57adaff31d14e0799698f3f2242fbd193f96cca51f751415e62c82941bfd4c"} Mar 09 09:59:56 crc kubenswrapper[4692]: I0309 09:59:56.091014 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" path="/var/lib/kubelet/pods/92f1158a-ddbf-4fd5-b446-c24f01cb3c66/volumes" Mar 09 09:59:56 crc kubenswrapper[4692]: I0309 09:59:56.926891 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:56 crc kubenswrapper[4692]: I0309 09:59:56.977450 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx"] Mar 09 09:59:56 crc kubenswrapper[4692]: I0309 09:59:56.986256 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx"] Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.066850 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-scripts\") pod \"bdc6dd32-00a7-4164-8f35-320700382b1f\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.066940 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-swiftconf\") pod \"bdc6dd32-00a7-4164-8f35-320700382b1f\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.066970 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-ring-data-devices\") pod \"bdc6dd32-00a7-4164-8f35-320700382b1f\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.067011 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-dispersionconf\") pod \"bdc6dd32-00a7-4164-8f35-320700382b1f\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.067050 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdc6dd32-00a7-4164-8f35-320700382b1f-etc-swift\") pod \"bdc6dd32-00a7-4164-8f35-320700382b1f\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.067182 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jknwj\" (UniqueName: \"kubernetes.io/projected/bdc6dd32-00a7-4164-8f35-320700382b1f-kube-api-access-jknwj\") pod \"bdc6dd32-00a7-4164-8f35-320700382b1f\" (UID: \"bdc6dd32-00a7-4164-8f35-320700382b1f\") " Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.067734 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bdc6dd32-00a7-4164-8f35-320700382b1f" (UID: "bdc6dd32-00a7-4164-8f35-320700382b1f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.068360 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdc6dd32-00a7-4164-8f35-320700382b1f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bdc6dd32-00a7-4164-8f35-320700382b1f" (UID: "bdc6dd32-00a7-4164-8f35-320700382b1f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.074005 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc6dd32-00a7-4164-8f35-320700382b1f-kube-api-access-jknwj" (OuterVolumeSpecName: "kube-api-access-jknwj") pod "bdc6dd32-00a7-4164-8f35-320700382b1f" (UID: "bdc6dd32-00a7-4164-8f35-320700382b1f"). InnerVolumeSpecName "kube-api-access-jknwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.089827 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-scripts" (OuterVolumeSpecName: "scripts") pod "bdc6dd32-00a7-4164-8f35-320700382b1f" (UID: "bdc6dd32-00a7-4164-8f35-320700382b1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.091823 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bdc6dd32-00a7-4164-8f35-320700382b1f" (UID: "bdc6dd32-00a7-4164-8f35-320700382b1f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.092768 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bdc6dd32-00a7-4164-8f35-320700382b1f" (UID: "bdc6dd32-00a7-4164-8f35-320700382b1f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.168873 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jknwj\" (UniqueName: \"kubernetes.io/projected/bdc6dd32-00a7-4164-8f35-320700382b1f-kube-api-access-jknwj\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.169241 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.169262 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.169275 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bdc6dd32-00a7-4164-8f35-320700382b1f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.169284 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bdc6dd32-00a7-4164-8f35-320700382b1f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.169294 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bdc6dd32-00a7-4164-8f35-320700382b1f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.607916 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d7b8581d07bd3ca6e1b1f21fc2842f76774c9535322b232feeb6af828051f3c" Mar 09 09:59:57 crc kubenswrapper[4692]: I0309 09:59:57.608035 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-qrxcx" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.079611 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc6dd32-00a7-4164-8f35-320700382b1f" path="/var/lib/kubelet/pods/bdc6dd32-00a7-4164-8f35-320700382b1f/volumes" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.136101 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z"] Mar 09 09:59:58 crc kubenswrapper[4692]: E0309 09:59:58.136438 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="registry-server" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.136451 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="registry-server" Mar 09 09:59:58 crc kubenswrapper[4692]: E0309 09:59:58.136465 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="extract-content" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.136472 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="extract-content" Mar 09 09:59:58 crc kubenswrapper[4692]: E0309 09:59:58.136482 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="extract-utilities" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.136487 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="extract-utilities" Mar 09 09:59:58 crc kubenswrapper[4692]: E0309 09:59:58.136515 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc6dd32-00a7-4164-8f35-320700382b1f" containerName="swift-ring-rebalance" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.136521 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc6dd32-00a7-4164-8f35-320700382b1f" containerName="swift-ring-rebalance" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.136657 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f1158a-ddbf-4fd5-b446-c24f01cb3c66" containerName="registry-server" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.136675 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc6dd32-00a7-4164-8f35-320700382b1f" containerName="swift-ring-rebalance" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.137137 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.139146 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.139374 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.150842 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z"] Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.287475 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-ring-data-devices\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.287547 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j5fk\" (UniqueName: \"kubernetes.io/projected/8eee70b9-3b2c-42dd-95cd-ec814120caf1-kube-api-access-2j5fk\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.287575 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8eee70b9-3b2c-42dd-95cd-ec814120caf1-etc-swift\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.287632 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-scripts\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.287775 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-dispersionconf\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.287903 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-swiftconf\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.389694 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-swiftconf\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.389778 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j5fk\" (UniqueName: \"kubernetes.io/projected/8eee70b9-3b2c-42dd-95cd-ec814120caf1-kube-api-access-2j5fk\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.389807 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-ring-data-devices\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.389833 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8eee70b9-3b2c-42dd-95cd-ec814120caf1-etc-swift\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.389892 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-scripts\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.389941 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-dispersionconf\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.391005 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8eee70b9-3b2c-42dd-95cd-ec814120caf1-etc-swift\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.391195 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-scripts\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.391209 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-ring-data-devices\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.399539 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-dispersionconf\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.402870 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-swiftconf\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.419131 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j5fk\" (UniqueName: \"kubernetes.io/projected/8eee70b9-3b2c-42dd-95cd-ec814120caf1-kube-api-access-2j5fk\") pod \"swift-ring-rebalance-debug-j7x7z\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.453661 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 09:59:58 crc kubenswrapper[4692]: I0309 09:59:58.765597 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z"] Mar 09 09:59:58 crc kubenswrapper[4692]: W0309 09:59:58.769422 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8eee70b9_3b2c_42dd_95cd_ec814120caf1.slice/crio-a1aa7874354c4cd9de03b7f862a99d792af3af70362d073eef35d9d7f03c0c5d WatchSource:0}: Error finding container a1aa7874354c4cd9de03b7f862a99d792af3af70362d073eef35d9d7f03c0c5d: Status 404 returned error can't find the container with id a1aa7874354c4cd9de03b7f862a99d792af3af70362d073eef35d9d7f03c0c5d Mar 09 09:59:59 crc kubenswrapper[4692]: I0309 09:59:59.631386 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" event={"ID":"8eee70b9-3b2c-42dd-95cd-ec814120caf1","Type":"ContainerStarted","Data":"40e18e32343d84552d4907a88b01d94aa799a0fa8c74f9d9b066d1c2154c2a70"} Mar 09 09:59:59 crc kubenswrapper[4692]: I0309 09:59:59.631821 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" event={"ID":"8eee70b9-3b2c-42dd-95cd-ec814120caf1","Type":"ContainerStarted","Data":"a1aa7874354c4cd9de03b7f862a99d792af3af70362d073eef35d9d7f03c0c5d"} Mar 09 09:59:59 crc kubenswrapper[4692]: I0309 09:59:59.657842 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" podStartSLOduration=1.657811303 podStartE2EDuration="1.657811303s" podCreationTimestamp="2026-03-09 09:59:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 09:59:59.650218548 +0000 UTC m=+2400.474954139" watchObservedRunningTime="2026-03-09 09:59:59.657811303 +0000 UTC m=+2400.482546894" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.160303 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g"] Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.161382 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.165507 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.165873 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.168270 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550840-tqnz9"] Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.170017 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.172246 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.172407 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.172619 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.190705 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g"] Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.200233 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550840-tqnz9"] Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.332313 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10f2c004-d6e9-4fa4-a694-65df002a3f72-config-volume\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.332389 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkvhb\" (UniqueName: \"kubernetes.io/projected/10f2c004-d6e9-4fa4-a694-65df002a3f72-kube-api-access-tkvhb\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.332827 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10f2c004-d6e9-4fa4-a694-65df002a3f72-secret-volume\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.333102 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d88tb\" (UniqueName: \"kubernetes.io/projected/ef13db35-bc96-4421-93a0-e8690b87d300-kube-api-access-d88tb\") pod \"auto-csr-approver-29550840-tqnz9\" (UID: \"ef13db35-bc96-4421-93a0-e8690b87d300\") " pod="openshift-infra/auto-csr-approver-29550840-tqnz9" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.434822 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10f2c004-d6e9-4fa4-a694-65df002a3f72-config-volume\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.435257 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkvhb\" (UniqueName: \"kubernetes.io/projected/10f2c004-d6e9-4fa4-a694-65df002a3f72-kube-api-access-tkvhb\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.435333 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10f2c004-d6e9-4fa4-a694-65df002a3f72-secret-volume\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.435406 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d88tb\" (UniqueName: \"kubernetes.io/projected/ef13db35-bc96-4421-93a0-e8690b87d300-kube-api-access-d88tb\") pod \"auto-csr-approver-29550840-tqnz9\" (UID: \"ef13db35-bc96-4421-93a0-e8690b87d300\") " pod="openshift-infra/auto-csr-approver-29550840-tqnz9" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.436092 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10f2c004-d6e9-4fa4-a694-65df002a3f72-config-volume\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.443467 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10f2c004-d6e9-4fa4-a694-65df002a3f72-secret-volume\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.453943 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkvhb\" (UniqueName: \"kubernetes.io/projected/10f2c004-d6e9-4fa4-a694-65df002a3f72-kube-api-access-tkvhb\") pod \"collect-profiles-29550840-6x64g\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.455275 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d88tb\" (UniqueName: \"kubernetes.io/projected/ef13db35-bc96-4421-93a0-e8690b87d300-kube-api-access-d88tb\") pod \"auto-csr-approver-29550840-tqnz9\" (UID: \"ef13db35-bc96-4421-93a0-e8690b87d300\") " pod="openshift-infra/auto-csr-approver-29550840-tqnz9" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.500882 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.521412 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.652127 4692 generic.go:334] "Generic (PLEG): container finished" podID="8eee70b9-3b2c-42dd-95cd-ec814120caf1" containerID="40e18e32343d84552d4907a88b01d94aa799a0fa8c74f9d9b066d1c2154c2a70" exitCode=0 Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.652219 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" event={"ID":"8eee70b9-3b2c-42dd-95cd-ec814120caf1","Type":"ContainerDied","Data":"40e18e32343d84552d4907a88b01d94aa799a0fa8c74f9d9b066d1c2154c2a70"} Mar 09 10:00:00 crc kubenswrapper[4692]: I0309 10:00:00.985716 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g"] Mar 09 10:00:01 crc kubenswrapper[4692]: I0309 10:00:01.054378 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550840-tqnz9"] Mar 09 10:00:01 crc kubenswrapper[4692]: W0309 10:00:01.063629 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef13db35_bc96_4421_93a0_e8690b87d300.slice/crio-b76907fc705ae1e1ac3a1c8d71b8512968c9cd5fd0888201cc713f7b4d276af9 WatchSource:0}: Error finding container b76907fc705ae1e1ac3a1c8d71b8512968c9cd5fd0888201cc713f7b4d276af9: Status 404 returned error can't find the container with id b76907fc705ae1e1ac3a1c8d71b8512968c9cd5fd0888201cc713f7b4d276af9 Mar 09 10:00:01 crc kubenswrapper[4692]: I0309 10:00:01.664764 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" event={"ID":"ef13db35-bc96-4421-93a0-e8690b87d300","Type":"ContainerStarted","Data":"b76907fc705ae1e1ac3a1c8d71b8512968c9cd5fd0888201cc713f7b4d276af9"} Mar 09 10:00:01 crc kubenswrapper[4692]: I0309 10:00:01.666783 4692 generic.go:334] "Generic (PLEG): container finished" podID="10f2c004-d6e9-4fa4-a694-65df002a3f72" containerID="26f288ba5ccd02d01a35a96ebea5592e1663715daf232f7c0f13035ef0565a0d" exitCode=0 Mar 09 10:00:01 crc kubenswrapper[4692]: I0309 10:00:01.666884 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" event={"ID":"10f2c004-d6e9-4fa4-a694-65df002a3f72","Type":"ContainerDied","Data":"26f288ba5ccd02d01a35a96ebea5592e1663715daf232f7c0f13035ef0565a0d"} Mar 09 10:00:01 crc kubenswrapper[4692]: I0309 10:00:01.666977 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" event={"ID":"10f2c004-d6e9-4fa4-a694-65df002a3f72","Type":"ContainerStarted","Data":"ae13846505a50ca17e8d4560b1c57c54bc623dc6beb8e887e04559035b2b5aa8"} Mar 09 10:00:01 crc kubenswrapper[4692]: I0309 10:00:01.969828 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.006212 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z"] Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.013466 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z"] Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.072558 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-ring-data-devices\") pod \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.072667 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-swiftconf\") pod \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.072700 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j5fk\" (UniqueName: \"kubernetes.io/projected/8eee70b9-3b2c-42dd-95cd-ec814120caf1-kube-api-access-2j5fk\") pod \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.072798 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-dispersionconf\") pod \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.072854 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8eee70b9-3b2c-42dd-95cd-ec814120caf1-etc-swift\") pod \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.072888 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-scripts\") pod \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\" (UID: \"8eee70b9-3b2c-42dd-95cd-ec814120caf1\") " Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.074389 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8eee70b9-3b2c-42dd-95cd-ec814120caf1-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "8eee70b9-3b2c-42dd-95cd-ec814120caf1" (UID: "8eee70b9-3b2c-42dd-95cd-ec814120caf1"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.074518 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "8eee70b9-3b2c-42dd-95cd-ec814120caf1" (UID: "8eee70b9-3b2c-42dd-95cd-ec814120caf1"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.081665 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eee70b9-3b2c-42dd-95cd-ec814120caf1-kube-api-access-2j5fk" (OuterVolumeSpecName: "kube-api-access-2j5fk") pod "8eee70b9-3b2c-42dd-95cd-ec814120caf1" (UID: "8eee70b9-3b2c-42dd-95cd-ec814120caf1"). InnerVolumeSpecName "kube-api-access-2j5fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.102927 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "8eee70b9-3b2c-42dd-95cd-ec814120caf1" (UID: "8eee70b9-3b2c-42dd-95cd-ec814120caf1"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.103254 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-scripts" (OuterVolumeSpecName: "scripts") pod "8eee70b9-3b2c-42dd-95cd-ec814120caf1" (UID: "8eee70b9-3b2c-42dd-95cd-ec814120caf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.107332 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "8eee70b9-3b2c-42dd-95cd-ec814120caf1" (UID: "8eee70b9-3b2c-42dd-95cd-ec814120caf1"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.178524 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.178578 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/8eee70b9-3b2c-42dd-95cd-ec814120caf1-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.178594 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.178612 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j5fk\" (UniqueName: \"kubernetes.io/projected/8eee70b9-3b2c-42dd-95cd-ec814120caf1-kube-api-access-2j5fk\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.178671 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/8eee70b9-3b2c-42dd-95cd-ec814120caf1-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.178777 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/8eee70b9-3b2c-42dd-95cd-ec814120caf1-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.677013 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-j7x7z" Mar 09 10:00:02 crc kubenswrapper[4692]: I0309 10:00:02.678205 4692 scope.go:117] "RemoveContainer" containerID="40e18e32343d84552d4907a88b01d94aa799a0fa8c74f9d9b066d1c2154c2a70" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.018880 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.093712 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10f2c004-d6e9-4fa4-a694-65df002a3f72-secret-volume\") pod \"10f2c004-d6e9-4fa4-a694-65df002a3f72\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.094037 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkvhb\" (UniqueName: \"kubernetes.io/projected/10f2c004-d6e9-4fa4-a694-65df002a3f72-kube-api-access-tkvhb\") pod \"10f2c004-d6e9-4fa4-a694-65df002a3f72\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.094133 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10f2c004-d6e9-4fa4-a694-65df002a3f72-config-volume\") pod \"10f2c004-d6e9-4fa4-a694-65df002a3f72\" (UID: \"10f2c004-d6e9-4fa4-a694-65df002a3f72\") " Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.095050 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10f2c004-d6e9-4fa4-a694-65df002a3f72-config-volume" (OuterVolumeSpecName: "config-volume") pod "10f2c004-d6e9-4fa4-a694-65df002a3f72" (UID: "10f2c004-d6e9-4fa4-a694-65df002a3f72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.097251 4692 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/10f2c004-d6e9-4fa4-a694-65df002a3f72-config-volume\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.102403 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f2c004-d6e9-4fa4-a694-65df002a3f72-kube-api-access-tkvhb" (OuterVolumeSpecName: "kube-api-access-tkvhb") pod "10f2c004-d6e9-4fa4-a694-65df002a3f72" (UID: "10f2c004-d6e9-4fa4-a694-65df002a3f72"). InnerVolumeSpecName "kube-api-access-tkvhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.109462 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f2c004-d6e9-4fa4-a694-65df002a3f72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "10f2c004-d6e9-4fa4-a694-65df002a3f72" (UID: "10f2c004-d6e9-4fa4-a694-65df002a3f72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.200112 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkvhb\" (UniqueName: \"kubernetes.io/projected/10f2c004-d6e9-4fa4-a694-65df002a3f72-kube-api-access-tkvhb\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.200181 4692 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/10f2c004-d6e9-4fa4-a694-65df002a3f72-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.206105 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khvqn"] Mar 09 10:00:03 crc kubenswrapper[4692]: E0309 10:00:03.206800 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eee70b9-3b2c-42dd-95cd-ec814120caf1" containerName="swift-ring-rebalance" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.207394 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eee70b9-3b2c-42dd-95cd-ec814120caf1" containerName="swift-ring-rebalance" Mar 09 10:00:03 crc kubenswrapper[4692]: E0309 10:00:03.207422 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f2c004-d6e9-4fa4-a694-65df002a3f72" containerName="collect-profiles" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.207433 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f2c004-d6e9-4fa4-a694-65df002a3f72" containerName="collect-profiles" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.215580 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eee70b9-3b2c-42dd-95cd-ec814120caf1" containerName="swift-ring-rebalance" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.215631 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f2c004-d6e9-4fa4-a694-65df002a3f72" containerName="collect-profiles" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.216570 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khvqn"] Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.216727 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.223769 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.224059 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.301684 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79a5ec34-efb6-4f55-86c4-48f79399df50-etc-swift\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.302122 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4j28\" (UniqueName: \"kubernetes.io/projected/79a5ec34-efb6-4f55-86c4-48f79399df50-kube-api-access-q4j28\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.302399 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-swiftconf\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.302552 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-ring-data-devices\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.302695 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-scripts\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.302826 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-dispersionconf\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.404090 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79a5ec34-efb6-4f55-86c4-48f79399df50-etc-swift\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.404192 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4j28\" (UniqueName: \"kubernetes.io/projected/79a5ec34-efb6-4f55-86c4-48f79399df50-kube-api-access-q4j28\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.404235 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-swiftconf\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.404274 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-ring-data-devices\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.404311 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-scripts\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.404332 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-dispersionconf\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.404723 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79a5ec34-efb6-4f55-86c4-48f79399df50-etc-swift\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.405459 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-ring-data-devices\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.405478 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-scripts\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.410765 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-swiftconf\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.410817 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-dispersionconf\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.424838 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4j28\" (UniqueName: \"kubernetes.io/projected/79a5ec34-efb6-4f55-86c4-48f79399df50-kube-api-access-q4j28\") pod \"swift-ring-rebalance-debug-khvqn\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.557546 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.696811 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" event={"ID":"10f2c004-d6e9-4fa4-a694-65df002a3f72","Type":"ContainerDied","Data":"ae13846505a50ca17e8d4560b1c57c54bc623dc6beb8e887e04559035b2b5aa8"} Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.696854 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae13846505a50ca17e8d4560b1c57c54bc623dc6beb8e887e04559035b2b5aa8" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.696929 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29550840-6x64g" Mar 09 10:00:03 crc kubenswrapper[4692]: I0309 10:00:03.838815 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khvqn"] Mar 09 10:00:03 crc kubenswrapper[4692]: W0309 10:00:03.851639 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79a5ec34_efb6_4f55_86c4_48f79399df50.slice/crio-53308b05ca77b53bb0485621f9d24216ffd6180fe54021c010a886c32d125e79 WatchSource:0}: Error finding container 53308b05ca77b53bb0485621f9d24216ffd6180fe54021c010a886c32d125e79: Status 404 returned error can't find the container with id 53308b05ca77b53bb0485621f9d24216ffd6180fe54021c010a886c32d125e79 Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.072339 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:00:04 crc kubenswrapper[4692]: E0309 10:00:04.072665 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.081439 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eee70b9-3b2c-42dd-95cd-ec814120caf1" path="/var/lib/kubelet/pods/8eee70b9-3b2c-42dd-95cd-ec814120caf1/volumes" Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.149617 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv"] Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.155970 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29550795-5rfpv"] Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.711695 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" event={"ID":"ef13db35-bc96-4421-93a0-e8690b87d300","Type":"ContainerStarted","Data":"b286711ebf13e74405f3c83ef50f29eb75fda7aaded659743ddae1074d870598"} Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.713715 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" event={"ID":"79a5ec34-efb6-4f55-86c4-48f79399df50","Type":"ContainerStarted","Data":"1ec9bcafe9f6d5c5553ec0dedfb27d9f211f302064fd33bfe663880e47c78589"} Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.713758 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" event={"ID":"79a5ec34-efb6-4f55-86c4-48f79399df50","Type":"ContainerStarted","Data":"53308b05ca77b53bb0485621f9d24216ffd6180fe54021c010a886c32d125e79"} Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.730297 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" podStartSLOduration=1.59889234 podStartE2EDuration="4.730267214s" podCreationTimestamp="2026-03-09 10:00:00 +0000 UTC" firstStartedPulling="2026-03-09 10:00:01.067315454 +0000 UTC m=+2401.892051035" lastFinishedPulling="2026-03-09 10:00:04.198690328 +0000 UTC m=+2405.023425909" observedRunningTime="2026-03-09 10:00:04.727802914 +0000 UTC m=+2405.552538505" watchObservedRunningTime="2026-03-09 10:00:04.730267214 +0000 UTC m=+2405.555002805" Mar 09 10:00:04 crc kubenswrapper[4692]: I0309 10:00:04.749860 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" podStartSLOduration=1.749831757 podStartE2EDuration="1.749831757s" podCreationTimestamp="2026-03-09 10:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:04.746670158 +0000 UTC m=+2405.571405759" watchObservedRunningTime="2026-03-09 10:00:04.749831757 +0000 UTC m=+2405.574567348" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.325552 4692 scope.go:117] "RemoveContainer" containerID="649278ae134347f516ec669ac93194c96cb0eebb54fb43b495abcd13e1bf6de1" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.372140 4692 scope.go:117] "RemoveContainer" containerID="869b74c8b7506c0252b5b0c7d192121a8c597138f6ebfd145a8875e9b4da5501" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.435891 4692 scope.go:117] "RemoveContainer" containerID="9770ea0091fa66c133afeb08906c31ff59c3bdc74f88c85deca0b7b375da682b" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.471980 4692 scope.go:117] "RemoveContainer" containerID="6e235a46dd97fdddeb3c18a7b1b4798244db0a5765b2f7df05b044d79da2a1b6" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.499884 4692 scope.go:117] "RemoveContainer" containerID="8f2a2cc5ed500ff11c2c0e98a8a512d7cb04e99f8d914e177f065f4fc4fb99f4" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.615737 4692 scope.go:117] "RemoveContainer" containerID="44adc8a44c7bc291cc756cf5cb875bb165f1af67c97281123ade2b3897125a9c" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.654558 4692 scope.go:117] "RemoveContainer" containerID="e5f0980ba2b99eca290cec57dc1e0c83f1bd0da6ffbcf57590d08929b8a48345" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.686070 4692 scope.go:117] "RemoveContainer" containerID="8a2e2289a00d1bb19f5449795b30f50a7cb8bb53977e98b02dad05c2006a8d74" Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.743309 4692 generic.go:334] "Generic (PLEG): container finished" podID="79a5ec34-efb6-4f55-86c4-48f79399df50" containerID="1ec9bcafe9f6d5c5553ec0dedfb27d9f211f302064fd33bfe663880e47c78589" exitCode=0 Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.744287 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" event={"ID":"79a5ec34-efb6-4f55-86c4-48f79399df50","Type":"ContainerDied","Data":"1ec9bcafe9f6d5c5553ec0dedfb27d9f211f302064fd33bfe663880e47c78589"} Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.763469 4692 generic.go:334] "Generic (PLEG): container finished" podID="ef13db35-bc96-4421-93a0-e8690b87d300" containerID="b286711ebf13e74405f3c83ef50f29eb75fda7aaded659743ddae1074d870598" exitCode=0 Mar 09 10:00:05 crc kubenswrapper[4692]: I0309 10:00:05.763600 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" event={"ID":"ef13db35-bc96-4421-93a0-e8690b87d300","Type":"ContainerDied","Data":"b286711ebf13e74405f3c83ef50f29eb75fda7aaded659743ddae1074d870598"} Mar 09 10:00:06 crc kubenswrapper[4692]: I0309 10:00:06.082227 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f96b54e5-804b-4d9e-9df9-538352c52073" path="/var/lib/kubelet/pods/f96b54e5-804b-4d9e-9df9-538352c52073/volumes" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.127554 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.135060 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.185366 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khvqn"] Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.186147 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79a5ec34-efb6-4f55-86c4-48f79399df50-etc-swift\") pod \"79a5ec34-efb6-4f55-86c4-48f79399df50\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.186383 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-swiftconf\") pod \"79a5ec34-efb6-4f55-86c4-48f79399df50\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.186426 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-scripts\") pod \"79a5ec34-efb6-4f55-86c4-48f79399df50\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.186468 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d88tb\" (UniqueName: \"kubernetes.io/projected/ef13db35-bc96-4421-93a0-e8690b87d300-kube-api-access-d88tb\") pod \"ef13db35-bc96-4421-93a0-e8690b87d300\" (UID: \"ef13db35-bc96-4421-93a0-e8690b87d300\") " Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.186500 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4j28\" (UniqueName: \"kubernetes.io/projected/79a5ec34-efb6-4f55-86c4-48f79399df50-kube-api-access-q4j28\") pod \"79a5ec34-efb6-4f55-86c4-48f79399df50\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.186519 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-ring-data-devices\") pod \"79a5ec34-efb6-4f55-86c4-48f79399df50\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.186556 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-dispersionconf\") pod \"79a5ec34-efb6-4f55-86c4-48f79399df50\" (UID: \"79a5ec34-efb6-4f55-86c4-48f79399df50\") " Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.188112 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79a5ec34-efb6-4f55-86c4-48f79399df50-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "79a5ec34-efb6-4f55-86c4-48f79399df50" (UID: "79a5ec34-efb6-4f55-86c4-48f79399df50"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.188096 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "79a5ec34-efb6-4f55-86c4-48f79399df50" (UID: "79a5ec34-efb6-4f55-86c4-48f79399df50"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.191661 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-khvqn"] Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.197659 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef13db35-bc96-4421-93a0-e8690b87d300-kube-api-access-d88tb" (OuterVolumeSpecName: "kube-api-access-d88tb") pod "ef13db35-bc96-4421-93a0-e8690b87d300" (UID: "ef13db35-bc96-4421-93a0-e8690b87d300"). InnerVolumeSpecName "kube-api-access-d88tb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.205548 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a5ec34-efb6-4f55-86c4-48f79399df50-kube-api-access-q4j28" (OuterVolumeSpecName: "kube-api-access-q4j28") pod "79a5ec34-efb6-4f55-86c4-48f79399df50" (UID: "79a5ec34-efb6-4f55-86c4-48f79399df50"). InnerVolumeSpecName "kube-api-access-q4j28". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.210274 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-scripts" (OuterVolumeSpecName: "scripts") pod "79a5ec34-efb6-4f55-86c4-48f79399df50" (UID: "79a5ec34-efb6-4f55-86c4-48f79399df50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.215339 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "79a5ec34-efb6-4f55-86c4-48f79399df50" (UID: "79a5ec34-efb6-4f55-86c4-48f79399df50"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.221495 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "79a5ec34-efb6-4f55-86c4-48f79399df50" (UID: "79a5ec34-efb6-4f55-86c4-48f79399df50"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.289428 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.289763 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/79a5ec34-efb6-4f55-86c4-48f79399df50-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.289843 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/79a5ec34-efb6-4f55-86c4-48f79399df50-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.289918 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.289998 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d88tb\" (UniqueName: \"kubernetes.io/projected/ef13db35-bc96-4421-93a0-e8690b87d300-kube-api-access-d88tb\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.290126 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4j28\" (UniqueName: \"kubernetes.io/projected/79a5ec34-efb6-4f55-86c4-48f79399df50-kube-api-access-q4j28\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.290232 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/79a5ec34-efb6-4f55-86c4-48f79399df50-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.785377 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550834-vbwpc"] Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.798427 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550834-vbwpc"] Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.803639 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" event={"ID":"ef13db35-bc96-4421-93a0-e8690b87d300","Type":"ContainerDied","Data":"b76907fc705ae1e1ac3a1c8d71b8512968c9cd5fd0888201cc713f7b4d276af9"} Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.803694 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b76907fc705ae1e1ac3a1c8d71b8512968c9cd5fd0888201cc713f7b4d276af9" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.803777 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550840-tqnz9" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.808998 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53308b05ca77b53bb0485621f9d24216ffd6180fe54021c010a886c32d125e79" Mar 09 10:00:07 crc kubenswrapper[4692]: I0309 10:00:07.809136 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-khvqn" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.082567 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a0af45c-6d41-42fe-aa77-c587ed20b0dc" path="/var/lib/kubelet/pods/5a0af45c-6d41-42fe-aa77-c587ed20b0dc/volumes" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.083694 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a5ec34-efb6-4f55-86c4-48f79399df50" path="/var/lib/kubelet/pods/79a5ec34-efb6-4f55-86c4-48f79399df50/volumes" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.361229 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rggj6"] Mar 09 10:00:08 crc kubenswrapper[4692]: E0309 10:00:08.361659 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79a5ec34-efb6-4f55-86c4-48f79399df50" containerName="swift-ring-rebalance" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.361679 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="79a5ec34-efb6-4f55-86c4-48f79399df50" containerName="swift-ring-rebalance" Mar 09 10:00:08 crc kubenswrapper[4692]: E0309 10:00:08.361693 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef13db35-bc96-4421-93a0-e8690b87d300" containerName="oc" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.361701 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef13db35-bc96-4421-93a0-e8690b87d300" containerName="oc" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.361878 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef13db35-bc96-4421-93a0-e8690b87d300" containerName="oc" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.361912 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="79a5ec34-efb6-4f55-86c4-48f79399df50" containerName="swift-ring-rebalance" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.362642 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.365013 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.366382 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.378481 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rggj6"] Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.409139 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-ring-data-devices\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.410070 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fc2j\" (UniqueName: \"kubernetes.io/projected/a715ded3-d99b-4cda-82a8-e0701e69fcec-kube-api-access-4fc2j\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.410221 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-swiftconf\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.410361 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-dispersionconf\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.410487 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a715ded3-d99b-4cda-82a8-e0701e69fcec-etc-swift\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.410659 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-scripts\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.512295 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-ring-data-devices\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.512358 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fc2j\" (UniqueName: \"kubernetes.io/projected/a715ded3-d99b-4cda-82a8-e0701e69fcec-kube-api-access-4fc2j\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.512385 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-swiftconf\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.512413 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-dispersionconf\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.512433 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a715ded3-d99b-4cda-82a8-e0701e69fcec-etc-swift\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.512474 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-scripts\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.513526 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-scripts\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.513763 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a715ded3-d99b-4cda-82a8-e0701e69fcec-etc-swift\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.513908 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-ring-data-devices\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.518113 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-dispersionconf\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.519481 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-swiftconf\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.532417 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fc2j\" (UniqueName: \"kubernetes.io/projected/a715ded3-d99b-4cda-82a8-e0701e69fcec-kube-api-access-4fc2j\") pod \"swift-ring-rebalance-debug-rggj6\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:08 crc kubenswrapper[4692]: I0309 10:00:08.682187 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:09 crc kubenswrapper[4692]: I0309 10:00:09.162293 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rggj6"] Mar 09 10:00:09 crc kubenswrapper[4692]: W0309 10:00:09.164197 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda715ded3_d99b_4cda_82a8_e0701e69fcec.slice/crio-f717e1a6d124d56cc02e7e94c3196996056d12317715b2747c33f4e21ddf16bf WatchSource:0}: Error finding container f717e1a6d124d56cc02e7e94c3196996056d12317715b2747c33f4e21ddf16bf: Status 404 returned error can't find the container with id f717e1a6d124d56cc02e7e94c3196996056d12317715b2747c33f4e21ddf16bf Mar 09 10:00:09 crc kubenswrapper[4692]: I0309 10:00:09.837335 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" event={"ID":"a715ded3-d99b-4cda-82a8-e0701e69fcec","Type":"ContainerStarted","Data":"45ddd12681fd658c3d4fa14b32797701a5d87cdb091b32dfb4b5a9e88a9eb338"} Mar 09 10:00:09 crc kubenswrapper[4692]: I0309 10:00:09.837727 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" event={"ID":"a715ded3-d99b-4cda-82a8-e0701e69fcec","Type":"ContainerStarted","Data":"f717e1a6d124d56cc02e7e94c3196996056d12317715b2747c33f4e21ddf16bf"} Mar 09 10:00:09 crc kubenswrapper[4692]: I0309 10:00:09.859864 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" podStartSLOduration=1.8598346810000002 podStartE2EDuration="1.859834681s" podCreationTimestamp="2026-03-09 10:00:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:09.856136206 +0000 UTC m=+2410.680871787" watchObservedRunningTime="2026-03-09 10:00:09.859834681 +0000 UTC m=+2410.684570312" Mar 09 10:00:10 crc kubenswrapper[4692]: I0309 10:00:10.850122 4692 generic.go:334] "Generic (PLEG): container finished" podID="a715ded3-d99b-4cda-82a8-e0701e69fcec" containerID="45ddd12681fd658c3d4fa14b32797701a5d87cdb091b32dfb4b5a9e88a9eb338" exitCode=0 Mar 09 10:00:10 crc kubenswrapper[4692]: I0309 10:00:10.850195 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" event={"ID":"a715ded3-d99b-4cda-82a8-e0701e69fcec","Type":"ContainerDied","Data":"45ddd12681fd658c3d4fa14b32797701a5d87cdb091b32dfb4b5a9e88a9eb338"} Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.158864 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.235316 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rggj6"] Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.245511 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-rggj6"] Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.277938 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-scripts\") pod \"a715ded3-d99b-4cda-82a8-e0701e69fcec\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.278092 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-ring-data-devices\") pod \"a715ded3-d99b-4cda-82a8-e0701e69fcec\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.278128 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a715ded3-d99b-4cda-82a8-e0701e69fcec-etc-swift\") pod \"a715ded3-d99b-4cda-82a8-e0701e69fcec\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.278203 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-dispersionconf\") pod \"a715ded3-d99b-4cda-82a8-e0701e69fcec\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.278241 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fc2j\" (UniqueName: \"kubernetes.io/projected/a715ded3-d99b-4cda-82a8-e0701e69fcec-kube-api-access-4fc2j\") pod \"a715ded3-d99b-4cda-82a8-e0701e69fcec\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.278322 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-swiftconf\") pod \"a715ded3-d99b-4cda-82a8-e0701e69fcec\" (UID: \"a715ded3-d99b-4cda-82a8-e0701e69fcec\") " Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.279032 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "a715ded3-d99b-4cda-82a8-e0701e69fcec" (UID: "a715ded3-d99b-4cda-82a8-e0701e69fcec"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.279514 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a715ded3-d99b-4cda-82a8-e0701e69fcec-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "a715ded3-d99b-4cda-82a8-e0701e69fcec" (UID: "a715ded3-d99b-4cda-82a8-e0701e69fcec"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.290584 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a715ded3-d99b-4cda-82a8-e0701e69fcec-kube-api-access-4fc2j" (OuterVolumeSpecName: "kube-api-access-4fc2j") pod "a715ded3-d99b-4cda-82a8-e0701e69fcec" (UID: "a715ded3-d99b-4cda-82a8-e0701e69fcec"). InnerVolumeSpecName "kube-api-access-4fc2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.300901 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-scripts" (OuterVolumeSpecName: "scripts") pod "a715ded3-d99b-4cda-82a8-e0701e69fcec" (UID: "a715ded3-d99b-4cda-82a8-e0701e69fcec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.308872 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "a715ded3-d99b-4cda-82a8-e0701e69fcec" (UID: "a715ded3-d99b-4cda-82a8-e0701e69fcec"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.309449 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "a715ded3-d99b-4cda-82a8-e0701e69fcec" (UID: "a715ded3-d99b-4cda-82a8-e0701e69fcec"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.380322 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.380410 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/a715ded3-d99b-4cda-82a8-e0701e69fcec-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.380428 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/a715ded3-d99b-4cda-82a8-e0701e69fcec-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.380440 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.380455 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fc2j\" (UniqueName: \"kubernetes.io/projected/a715ded3-d99b-4cda-82a8-e0701e69fcec-kube-api-access-4fc2j\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.380466 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/a715ded3-d99b-4cda-82a8-e0701e69fcec-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.872837 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-rggj6" Mar 09 10:00:12 crc kubenswrapper[4692]: I0309 10:00:12.872689 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f717e1a6d124d56cc02e7e94c3196996056d12317715b2747c33f4e21ddf16bf" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.381540 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6"] Mar 09 10:00:13 crc kubenswrapper[4692]: E0309 10:00:13.382317 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a715ded3-d99b-4cda-82a8-e0701e69fcec" containerName="swift-ring-rebalance" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.382332 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="a715ded3-d99b-4cda-82a8-e0701e69fcec" containerName="swift-ring-rebalance" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.382484 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="a715ded3-d99b-4cda-82a8-e0701e69fcec" containerName="swift-ring-rebalance" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.383031 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.385879 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.386794 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.393002 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6"] Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.500364 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-swiftconf\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.500470 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-scripts\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.500504 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57jvr\" (UniqueName: \"kubernetes.io/projected/44923da2-e109-46a2-baad-b122034c53ef-kube-api-access-57jvr\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.500568 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-dispersionconf\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.500608 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-ring-data-devices\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.500680 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44923da2-e109-46a2-baad-b122034c53ef-etc-swift\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.601614 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44923da2-e109-46a2-baad-b122034c53ef-etc-swift\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.601693 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-swiftconf\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.601766 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-scripts\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.601794 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57jvr\" (UniqueName: \"kubernetes.io/projected/44923da2-e109-46a2-baad-b122034c53ef-kube-api-access-57jvr\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.601821 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-dispersionconf\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.601846 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-ring-data-devices\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.602145 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44923da2-e109-46a2-baad-b122034c53ef-etc-swift\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.602701 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-ring-data-devices\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.602802 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-scripts\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.608403 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-swiftconf\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.613704 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-dispersionconf\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.628720 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57jvr\" (UniqueName: \"kubernetes.io/projected/44923da2-e109-46a2-baad-b122034c53ef-kube-api-access-57jvr\") pod \"swift-ring-rebalance-debug-8kfg6\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:13 crc kubenswrapper[4692]: I0309 10:00:13.701738 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:14 crc kubenswrapper[4692]: I0309 10:00:14.089332 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a715ded3-d99b-4cda-82a8-e0701e69fcec" path="/var/lib/kubelet/pods/a715ded3-d99b-4cda-82a8-e0701e69fcec/volumes" Mar 09 10:00:14 crc kubenswrapper[4692]: I0309 10:00:14.149399 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6"] Mar 09 10:00:14 crc kubenswrapper[4692]: I0309 10:00:14.891911 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" event={"ID":"44923da2-e109-46a2-baad-b122034c53ef","Type":"ContainerStarted","Data":"ac4efa8079008baaf554c3ede0ef2424c4e673d99d05f101f3e6c486864f1d27"} Mar 09 10:00:14 crc kubenswrapper[4692]: I0309 10:00:14.892459 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" event={"ID":"44923da2-e109-46a2-baad-b122034c53ef","Type":"ContainerStarted","Data":"dfd919f16ba2be299f143e0fd1b0c99bda79125d6645a6d3e336bb999959a753"} Mar 09 10:00:15 crc kubenswrapper[4692]: I0309 10:00:15.071560 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:00:15 crc kubenswrapper[4692]: E0309 10:00:15.071807 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:00:15 crc kubenswrapper[4692]: I0309 10:00:15.907341 4692 generic.go:334] "Generic (PLEG): container finished" podID="44923da2-e109-46a2-baad-b122034c53ef" containerID="ac4efa8079008baaf554c3ede0ef2424c4e673d99d05f101f3e6c486864f1d27" exitCode=0 Mar 09 10:00:15 crc kubenswrapper[4692]: I0309 10:00:15.907413 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" event={"ID":"44923da2-e109-46a2-baad-b122034c53ef","Type":"ContainerDied","Data":"ac4efa8079008baaf554c3ede0ef2424c4e673d99d05f101f3e6c486864f1d27"} Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.188769 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.231404 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6"] Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.240481 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6"] Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.272871 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57jvr\" (UniqueName: \"kubernetes.io/projected/44923da2-e109-46a2-baad-b122034c53ef-kube-api-access-57jvr\") pod \"44923da2-e109-46a2-baad-b122034c53ef\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.273049 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-dispersionconf\") pod \"44923da2-e109-46a2-baad-b122034c53ef\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.273088 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-swiftconf\") pod \"44923da2-e109-46a2-baad-b122034c53ef\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.273414 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-scripts\") pod \"44923da2-e109-46a2-baad-b122034c53ef\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.273465 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44923da2-e109-46a2-baad-b122034c53ef-etc-swift\") pod \"44923da2-e109-46a2-baad-b122034c53ef\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.273514 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-ring-data-devices\") pod \"44923da2-e109-46a2-baad-b122034c53ef\" (UID: \"44923da2-e109-46a2-baad-b122034c53ef\") " Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.274594 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "44923da2-e109-46a2-baad-b122034c53ef" (UID: "44923da2-e109-46a2-baad-b122034c53ef"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.275688 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44923da2-e109-46a2-baad-b122034c53ef-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "44923da2-e109-46a2-baad-b122034c53ef" (UID: "44923da2-e109-46a2-baad-b122034c53ef"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.285447 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44923da2-e109-46a2-baad-b122034c53ef-kube-api-access-57jvr" (OuterVolumeSpecName: "kube-api-access-57jvr") pod "44923da2-e109-46a2-baad-b122034c53ef" (UID: "44923da2-e109-46a2-baad-b122034c53ef"). InnerVolumeSpecName "kube-api-access-57jvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.298121 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-scripts" (OuterVolumeSpecName: "scripts") pod "44923da2-e109-46a2-baad-b122034c53ef" (UID: "44923da2-e109-46a2-baad-b122034c53ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.305293 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "44923da2-e109-46a2-baad-b122034c53ef" (UID: "44923da2-e109-46a2-baad-b122034c53ef"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.310860 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "44923da2-e109-46a2-baad-b122034c53ef" (UID: "44923da2-e109-46a2-baad-b122034c53ef"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.376118 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.376198 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/44923da2-e109-46a2-baad-b122034c53ef-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.376219 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/44923da2-e109-46a2-baad-b122034c53ef-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.376243 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57jvr\" (UniqueName: \"kubernetes.io/projected/44923da2-e109-46a2-baad-b122034c53ef-kube-api-access-57jvr\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.376263 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.376280 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/44923da2-e109-46a2-baad-b122034c53ef-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.928007 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfd919f16ba2be299f143e0fd1b0c99bda79125d6645a6d3e336bb999959a753" Mar 09 10:00:17 crc kubenswrapper[4692]: I0309 10:00:17.928088 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8kfg6" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.081411 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44923da2-e109-46a2-baad-b122034c53ef" path="/var/lib/kubelet/pods/44923da2-e109-46a2-baad-b122034c53ef/volumes" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.402412 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg"] Mar 09 10:00:18 crc kubenswrapper[4692]: E0309 10:00:18.402707 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44923da2-e109-46a2-baad-b122034c53ef" containerName="swift-ring-rebalance" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.402719 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="44923da2-e109-46a2-baad-b122034c53ef" containerName="swift-ring-rebalance" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.402879 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="44923da2-e109-46a2-baad-b122034c53ef" containerName="swift-ring-rebalance" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.403428 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.405834 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.406406 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.414333 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg"] Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.493218 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-swiftconf\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.493304 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/458453d9-ea99-4ff5-bf3a-ccc71c141209-etc-swift\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.493342 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-ring-data-devices\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.493371 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-dispersionconf\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.493395 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-scripts\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.493414 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqswh\" (UniqueName: \"kubernetes.io/projected/458453d9-ea99-4ff5-bf3a-ccc71c141209-kube-api-access-sqswh\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.594995 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-ring-data-devices\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.595072 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-dispersionconf\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.595102 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-scripts\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.595119 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqswh\" (UniqueName: \"kubernetes.io/projected/458453d9-ea99-4ff5-bf3a-ccc71c141209-kube-api-access-sqswh\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.595195 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-swiftconf\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.595240 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/458453d9-ea99-4ff5-bf3a-ccc71c141209-etc-swift\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.595846 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/458453d9-ea99-4ff5-bf3a-ccc71c141209-etc-swift\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.595960 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-ring-data-devices\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.596191 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-scripts\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.600799 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-dispersionconf\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.603036 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-swiftconf\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.613949 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqswh\" (UniqueName: \"kubernetes.io/projected/458453d9-ea99-4ff5-bf3a-ccc71c141209-kube-api-access-sqswh\") pod \"swift-ring-rebalance-debug-7b9zg\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:18 crc kubenswrapper[4692]: I0309 10:00:18.720852 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:19 crc kubenswrapper[4692]: I0309 10:00:19.166748 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg"] Mar 09 10:00:19 crc kubenswrapper[4692]: I0309 10:00:19.965697 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" event={"ID":"458453d9-ea99-4ff5-bf3a-ccc71c141209","Type":"ContainerStarted","Data":"e9755eeb3813eec41173bf96202e4a5a090bbaba593bd94d74363956398a3064"} Mar 09 10:00:19 crc kubenswrapper[4692]: I0309 10:00:19.966243 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" event={"ID":"458453d9-ea99-4ff5-bf3a-ccc71c141209","Type":"ContainerStarted","Data":"83915be11042a3abd1ae09857917de36db51954c4d2dcaa45b7a0950eb26e57a"} Mar 09 10:00:19 crc kubenswrapper[4692]: I0309 10:00:19.987299 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" podStartSLOduration=1.98726984 podStartE2EDuration="1.98726984s" podCreationTimestamp="2026-03-09 10:00:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:19.983899275 +0000 UTC m=+2420.808634856" watchObservedRunningTime="2026-03-09 10:00:19.98726984 +0000 UTC m=+2420.812005421" Mar 09 10:00:20 crc kubenswrapper[4692]: I0309 10:00:20.978391 4692 generic.go:334] "Generic (PLEG): container finished" podID="458453d9-ea99-4ff5-bf3a-ccc71c141209" containerID="e9755eeb3813eec41173bf96202e4a5a090bbaba593bd94d74363956398a3064" exitCode=0 Mar 09 10:00:20 crc kubenswrapper[4692]: I0309 10:00:20.978474 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" event={"ID":"458453d9-ea99-4ff5-bf3a-ccc71c141209","Type":"ContainerDied","Data":"e9755eeb3813eec41173bf96202e4a5a090bbaba593bd94d74363956398a3064"} Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.359123 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.419377 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg"] Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.427814 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg"] Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.461471 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/458453d9-ea99-4ff5-bf3a-ccc71c141209-etc-swift\") pod \"458453d9-ea99-4ff5-bf3a-ccc71c141209\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.461559 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqswh\" (UniqueName: \"kubernetes.io/projected/458453d9-ea99-4ff5-bf3a-ccc71c141209-kube-api-access-sqswh\") pod \"458453d9-ea99-4ff5-bf3a-ccc71c141209\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.462597 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/458453d9-ea99-4ff5-bf3a-ccc71c141209-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "458453d9-ea99-4ff5-bf3a-ccc71c141209" (UID: "458453d9-ea99-4ff5-bf3a-ccc71c141209"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.462683 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-swiftconf\") pod \"458453d9-ea99-4ff5-bf3a-ccc71c141209\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.462769 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-ring-data-devices\") pod \"458453d9-ea99-4ff5-bf3a-ccc71c141209\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.462830 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-scripts\") pod \"458453d9-ea99-4ff5-bf3a-ccc71c141209\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.462945 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-dispersionconf\") pod \"458453d9-ea99-4ff5-bf3a-ccc71c141209\" (UID: \"458453d9-ea99-4ff5-bf3a-ccc71c141209\") " Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.463454 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "458453d9-ea99-4ff5-bf3a-ccc71c141209" (UID: "458453d9-ea99-4ff5-bf3a-ccc71c141209"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.463836 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.463868 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/458453d9-ea99-4ff5-bf3a-ccc71c141209-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.469621 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458453d9-ea99-4ff5-bf3a-ccc71c141209-kube-api-access-sqswh" (OuterVolumeSpecName: "kube-api-access-sqswh") pod "458453d9-ea99-4ff5-bf3a-ccc71c141209" (UID: "458453d9-ea99-4ff5-bf3a-ccc71c141209"). InnerVolumeSpecName "kube-api-access-sqswh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.493555 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "458453d9-ea99-4ff5-bf3a-ccc71c141209" (UID: "458453d9-ea99-4ff5-bf3a-ccc71c141209"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.496224 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-scripts" (OuterVolumeSpecName: "scripts") pod "458453d9-ea99-4ff5-bf3a-ccc71c141209" (UID: "458453d9-ea99-4ff5-bf3a-ccc71c141209"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.504423 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "458453d9-ea99-4ff5-bf3a-ccc71c141209" (UID: "458453d9-ea99-4ff5-bf3a-ccc71c141209"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.564776 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.564812 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqswh\" (UniqueName: \"kubernetes.io/projected/458453d9-ea99-4ff5-bf3a-ccc71c141209-kube-api-access-sqswh\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.564823 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/458453d9-ea99-4ff5-bf3a-ccc71c141209-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:22 crc kubenswrapper[4692]: I0309 10:00:22.564831 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/458453d9-ea99-4ff5-bf3a-ccc71c141209-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.000636 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83915be11042a3abd1ae09857917de36db51954c4d2dcaa45b7a0950eb26e57a" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.000692 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-7b9zg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.571428 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-crqkg"] Mar 09 10:00:23 crc kubenswrapper[4692]: E0309 10:00:23.571803 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458453d9-ea99-4ff5-bf3a-ccc71c141209" containerName="swift-ring-rebalance" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.571821 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="458453d9-ea99-4ff5-bf3a-ccc71c141209" containerName="swift-ring-rebalance" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.571990 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="458453d9-ea99-4ff5-bf3a-ccc71c141209" containerName="swift-ring-rebalance" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.572646 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.575439 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.575656 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.583764 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-crqkg"] Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.685099 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-ring-data-devices\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.685680 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-dispersionconf\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.685763 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2h92\" (UniqueName: \"kubernetes.io/projected/da8ea7d5-c48f-4353-a08d-1abfb1610abf-kube-api-access-f2h92\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.686066 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da8ea7d5-c48f-4353-a08d-1abfb1610abf-etc-swift\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.686198 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.686452 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-scripts\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.788339 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da8ea7d5-c48f-4353-a08d-1abfb1610abf-etc-swift\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.788392 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.788428 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-scripts\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.788467 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-ring-data-devices\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.788490 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-dispersionconf\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.788514 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2h92\" (UniqueName: \"kubernetes.io/projected/da8ea7d5-c48f-4353-a08d-1abfb1610abf-kube-api-access-f2h92\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.789768 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-ring-data-devices\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.790293 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-scripts\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.790726 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da8ea7d5-c48f-4353-a08d-1abfb1610abf-etc-swift\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.798061 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.800693 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-dispersionconf\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.809567 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2h92\" (UniqueName: \"kubernetes.io/projected/da8ea7d5-c48f-4353-a08d-1abfb1610abf-kube-api-access-f2h92\") pod \"swift-ring-rebalance-debug-crqkg\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:23 crc kubenswrapper[4692]: I0309 10:00:23.951717 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:24 crc kubenswrapper[4692]: I0309 10:00:24.090576 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="458453d9-ea99-4ff5-bf3a-ccc71c141209" path="/var/lib/kubelet/pods/458453d9-ea99-4ff5-bf3a-ccc71c141209/volumes" Mar 09 10:00:24 crc kubenswrapper[4692]: I0309 10:00:24.481118 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-crqkg"] Mar 09 10:00:25 crc kubenswrapper[4692]: I0309 10:00:25.027918 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" event={"ID":"da8ea7d5-c48f-4353-a08d-1abfb1610abf","Type":"ContainerStarted","Data":"db53439628543717a86a7cf97a9bafae4fecd55f1573c6554af0ea00935d73a1"} Mar 09 10:00:25 crc kubenswrapper[4692]: I0309 10:00:25.028378 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" event={"ID":"da8ea7d5-c48f-4353-a08d-1abfb1610abf","Type":"ContainerStarted","Data":"634d757f2bfb6d841ba7c8fc2c9a57bef38ce75ccf00246f00616581e81d306a"} Mar 09 10:00:25 crc kubenswrapper[4692]: I0309 10:00:25.060063 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" podStartSLOduration=2.060034922 podStartE2EDuration="2.060034922s" podCreationTimestamp="2026-03-09 10:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:25.050886904 +0000 UTC m=+2425.875622515" watchObservedRunningTime="2026-03-09 10:00:25.060034922 +0000 UTC m=+2425.884770513" Mar 09 10:00:27 crc kubenswrapper[4692]: I0309 10:00:27.047753 4692 generic.go:334] "Generic (PLEG): container finished" podID="da8ea7d5-c48f-4353-a08d-1abfb1610abf" containerID="db53439628543717a86a7cf97a9bafae4fecd55f1573c6554af0ea00935d73a1" exitCode=0 Mar 09 10:00:27 crc kubenswrapper[4692]: I0309 10:00:27.047848 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" event={"ID":"da8ea7d5-c48f-4353-a08d-1abfb1610abf","Type":"ContainerDied","Data":"db53439628543717a86a7cf97a9bafae4fecd55f1573c6554af0ea00935d73a1"} Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.362740 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.401442 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-crqkg"] Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.412326 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-crqkg"] Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.479936 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da8ea7d5-c48f-4353-a08d-1abfb1610abf-etc-swift\") pod \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.480416 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-dispersionconf\") pod \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.480551 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-scripts\") pod \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.480688 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2h92\" (UniqueName: \"kubernetes.io/projected/da8ea7d5-c48f-4353-a08d-1abfb1610abf-kube-api-access-f2h92\") pod \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.480894 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-ring-data-devices\") pod \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.481002 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf\") pod \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.481505 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "da8ea7d5-c48f-4353-a08d-1abfb1610abf" (UID: "da8ea7d5-c48f-4353-a08d-1abfb1610abf"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.481522 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da8ea7d5-c48f-4353-a08d-1abfb1610abf-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "da8ea7d5-c48f-4353-a08d-1abfb1610abf" (UID: "da8ea7d5-c48f-4353-a08d-1abfb1610abf"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.481754 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/da8ea7d5-c48f-4353-a08d-1abfb1610abf-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.481782 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.488499 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da8ea7d5-c48f-4353-a08d-1abfb1610abf-kube-api-access-f2h92" (OuterVolumeSpecName: "kube-api-access-f2h92") pod "da8ea7d5-c48f-4353-a08d-1abfb1610abf" (UID: "da8ea7d5-c48f-4353-a08d-1abfb1610abf"). InnerVolumeSpecName "kube-api-access-f2h92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:28 crc kubenswrapper[4692]: E0309 10:00:28.507571 4692 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf podName:da8ea7d5-c48f-4353-a08d-1abfb1610abf nodeName:}" failed. No retries permitted until 2026-03-09 10:00:29.007528582 +0000 UTC m=+2429.832264173 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "swiftconf" (UniqueName: "kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf") pod "da8ea7d5-c48f-4353-a08d-1abfb1610abf" (UID: "da8ea7d5-c48f-4353-a08d-1abfb1610abf") : error deleting /var/lib/kubelet/pods/da8ea7d5-c48f-4353-a08d-1abfb1610abf/volume-subpaths: remove /var/lib/kubelet/pods/da8ea7d5-c48f-4353-a08d-1abfb1610abf/volume-subpaths: no such file or directory Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.508296 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-scripts" (OuterVolumeSpecName: "scripts") pod "da8ea7d5-c48f-4353-a08d-1abfb1610abf" (UID: "da8ea7d5-c48f-4353-a08d-1abfb1610abf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.508318 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "da8ea7d5-c48f-4353-a08d-1abfb1610abf" (UID: "da8ea7d5-c48f-4353-a08d-1abfb1610abf"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.583078 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.583119 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da8ea7d5-c48f-4353-a08d-1abfb1610abf-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:28 crc kubenswrapper[4692]: I0309 10:00:28.583130 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2h92\" (UniqueName: \"kubernetes.io/projected/da8ea7d5-c48f-4353-a08d-1abfb1610abf-kube-api-access-f2h92\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.070858 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="634d757f2bfb6d841ba7c8fc2c9a57bef38ce75ccf00246f00616581e81d306a" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.071229 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-crqkg" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.092489 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf\") pod \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\" (UID: \"da8ea7d5-c48f-4353-a08d-1abfb1610abf\") " Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.097496 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "da8ea7d5-c48f-4353-a08d-1abfb1610abf" (UID: "da8ea7d5-c48f-4353-a08d-1abfb1610abf"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.194942 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/da8ea7d5-c48f-4353-a08d-1abfb1610abf-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.550071 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-57hdt"] Mar 09 10:00:29 crc kubenswrapper[4692]: E0309 10:00:29.550565 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da8ea7d5-c48f-4353-a08d-1abfb1610abf" containerName="swift-ring-rebalance" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.550582 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="da8ea7d5-c48f-4353-a08d-1abfb1610abf" containerName="swift-ring-rebalance" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.550773 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="da8ea7d5-c48f-4353-a08d-1abfb1610abf" containerName="swift-ring-rebalance" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.551465 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.553777 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.553854 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.571897 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-57hdt"] Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.703506 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-scripts\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.703587 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-ring-data-devices\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.703792 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/496330ec-95c0-4d28-9ce5-b90cf3a9805f-etc-swift\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.703967 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-swiftconf\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.704087 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfvr2\" (UniqueName: \"kubernetes.io/projected/496330ec-95c0-4d28-9ce5-b90cf3a9805f-kube-api-access-tfvr2\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.704181 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-dispersionconf\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.805502 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-ring-data-devices\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.805584 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/496330ec-95c0-4d28-9ce5-b90cf3a9805f-etc-swift\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.805608 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-swiftconf\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.805640 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfvr2\" (UniqueName: \"kubernetes.io/projected/496330ec-95c0-4d28-9ce5-b90cf3a9805f-kube-api-access-tfvr2\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.805669 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-dispersionconf\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.805715 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-scripts\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.806442 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/496330ec-95c0-4d28-9ce5-b90cf3a9805f-etc-swift\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.806812 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-scripts\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.806812 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-ring-data-devices\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.810816 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-dispersionconf\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.812616 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-swiftconf\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.823944 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfvr2\" (UniqueName: \"kubernetes.io/projected/496330ec-95c0-4d28-9ce5-b90cf3a9805f-kube-api-access-tfvr2\") pod \"swift-ring-rebalance-debug-57hdt\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:29 crc kubenswrapper[4692]: I0309 10:00:29.876527 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:30 crc kubenswrapper[4692]: I0309 10:00:30.083659 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:00:30 crc kubenswrapper[4692]: E0309 10:00:30.084309 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:00:30 crc kubenswrapper[4692]: I0309 10:00:30.087837 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da8ea7d5-c48f-4353-a08d-1abfb1610abf" path="/var/lib/kubelet/pods/da8ea7d5-c48f-4353-a08d-1abfb1610abf/volumes" Mar 09 10:00:30 crc kubenswrapper[4692]: I0309 10:00:30.384377 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-57hdt"] Mar 09 10:00:31 crc kubenswrapper[4692]: I0309 10:00:31.093800 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" event={"ID":"496330ec-95c0-4d28-9ce5-b90cf3a9805f","Type":"ContainerStarted","Data":"23dd09656a52382798ccf3887d78903320e412c04837135dae56a4e336d51d55"} Mar 09 10:00:31 crc kubenswrapper[4692]: I0309 10:00:31.094340 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" event={"ID":"496330ec-95c0-4d28-9ce5-b90cf3a9805f","Type":"ContainerStarted","Data":"5565fddd59566ff39b94793ca9cc4f9761a053c1972eb44c72ed771eb18af010"} Mar 09 10:00:31 crc kubenswrapper[4692]: I0309 10:00:31.122665 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" podStartSLOduration=2.122640683 podStartE2EDuration="2.122640683s" podCreationTimestamp="2026-03-09 10:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:31.119958647 +0000 UTC m=+2431.944694228" watchObservedRunningTime="2026-03-09 10:00:31.122640683 +0000 UTC m=+2431.947376274" Mar 09 10:00:33 crc kubenswrapper[4692]: I0309 10:00:33.124961 4692 generic.go:334] "Generic (PLEG): container finished" podID="496330ec-95c0-4d28-9ce5-b90cf3a9805f" containerID="23dd09656a52382798ccf3887d78903320e412c04837135dae56a4e336d51d55" exitCode=0 Mar 09 10:00:33 crc kubenswrapper[4692]: I0309 10:00:33.125062 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" event={"ID":"496330ec-95c0-4d28-9ce5-b90cf3a9805f","Type":"ContainerDied","Data":"23dd09656a52382798ccf3887d78903320e412c04837135dae56a4e336d51d55"} Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.554284 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.595412 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-57hdt"] Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.603746 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-57hdt"] Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.694875 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-swiftconf\") pod \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.695005 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfvr2\" (UniqueName: \"kubernetes.io/projected/496330ec-95c0-4d28-9ce5-b90cf3a9805f-kube-api-access-tfvr2\") pod \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.695042 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-dispersionconf\") pod \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.695215 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/496330ec-95c0-4d28-9ce5-b90cf3a9805f-etc-swift\") pod \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.695277 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-scripts\") pod \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.695311 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-ring-data-devices\") pod \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\" (UID: \"496330ec-95c0-4d28-9ce5-b90cf3a9805f\") " Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.696122 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "496330ec-95c0-4d28-9ce5-b90cf3a9805f" (UID: "496330ec-95c0-4d28-9ce5-b90cf3a9805f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.696378 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/496330ec-95c0-4d28-9ce5-b90cf3a9805f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "496330ec-95c0-4d28-9ce5-b90cf3a9805f" (UID: "496330ec-95c0-4d28-9ce5-b90cf3a9805f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.701976 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496330ec-95c0-4d28-9ce5-b90cf3a9805f-kube-api-access-tfvr2" (OuterVolumeSpecName: "kube-api-access-tfvr2") pod "496330ec-95c0-4d28-9ce5-b90cf3a9805f" (UID: "496330ec-95c0-4d28-9ce5-b90cf3a9805f"). InnerVolumeSpecName "kube-api-access-tfvr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.721427 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "496330ec-95c0-4d28-9ce5-b90cf3a9805f" (UID: "496330ec-95c0-4d28-9ce5-b90cf3a9805f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.725606 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "496330ec-95c0-4d28-9ce5-b90cf3a9805f" (UID: "496330ec-95c0-4d28-9ce5-b90cf3a9805f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.730137 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-scripts" (OuterVolumeSpecName: "scripts") pod "496330ec-95c0-4d28-9ce5-b90cf3a9805f" (UID: "496330ec-95c0-4d28-9ce5-b90cf3a9805f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.797517 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.797852 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfvr2\" (UniqueName: \"kubernetes.io/projected/496330ec-95c0-4d28-9ce5-b90cf3a9805f-kube-api-access-tfvr2\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.797959 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/496330ec-95c0-4d28-9ce5-b90cf3a9805f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.798050 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/496330ec-95c0-4d28-9ce5-b90cf3a9805f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.798133 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:34 crc kubenswrapper[4692]: I0309 10:00:34.798238 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/496330ec-95c0-4d28-9ce5-b90cf3a9805f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.144919 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5565fddd59566ff39b94793ca9cc4f9761a053c1972eb44c72ed771eb18af010" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.144984 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-57hdt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.752091 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt"] Mar 09 10:00:35 crc kubenswrapper[4692]: E0309 10:00:35.752495 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496330ec-95c0-4d28-9ce5-b90cf3a9805f" containerName="swift-ring-rebalance" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.752510 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="496330ec-95c0-4d28-9ce5-b90cf3a9805f" containerName="swift-ring-rebalance" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.752681 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="496330ec-95c0-4d28-9ce5-b90cf3a9805f" containerName="swift-ring-rebalance" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.753350 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.757004 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.757004 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.768370 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt"] Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.815296 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-ring-data-devices\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.815363 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-scripts\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.815409 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-dispersionconf\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.815463 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-etc-swift\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.815522 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl8rw\" (UniqueName: \"kubernetes.io/projected/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-kube-api-access-wl8rw\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.815549 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-swiftconf\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.916722 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl8rw\" (UniqueName: \"kubernetes.io/projected/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-kube-api-access-wl8rw\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.917100 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-swiftconf\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.917200 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-ring-data-devices\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.917312 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-scripts\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.917406 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-dispersionconf\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.917509 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-etc-swift\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.918460 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-etc-swift\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.918668 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-scripts\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.918678 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-ring-data-devices\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.925396 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-dispersionconf\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.925445 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-swiftconf\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:35 crc kubenswrapper[4692]: I0309 10:00:35.938208 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl8rw\" (UniqueName: \"kubernetes.io/projected/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-kube-api-access-wl8rw\") pod \"swift-ring-rebalance-debug-wtzqt\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:36 crc kubenswrapper[4692]: I0309 10:00:36.090379 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496330ec-95c0-4d28-9ce5-b90cf3a9805f" path="/var/lib/kubelet/pods/496330ec-95c0-4d28-9ce5-b90cf3a9805f/volumes" Mar 09 10:00:36 crc kubenswrapper[4692]: I0309 10:00:36.125702 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:36 crc kubenswrapper[4692]: I0309 10:00:36.597026 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt"] Mar 09 10:00:37 crc kubenswrapper[4692]: I0309 10:00:37.167576 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" event={"ID":"2123c2c0-3966-4550-a5c1-9ece6e4d1e67","Type":"ContainerStarted","Data":"d5124acfaebcf56097f649b0f4498f5d7cccf9b3b630932cef47b05fa31c840a"} Mar 09 10:00:37 crc kubenswrapper[4692]: I0309 10:00:37.167641 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" event={"ID":"2123c2c0-3966-4550-a5c1-9ece6e4d1e67","Type":"ContainerStarted","Data":"60bad8ca48fa404add901227bcfdbe80d01d58ee537dda64840df315fb7cdf60"} Mar 09 10:00:37 crc kubenswrapper[4692]: I0309 10:00:37.190241 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" podStartSLOduration=2.190215453 podStartE2EDuration="2.190215453s" podCreationTimestamp="2026-03-09 10:00:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:37.184596094 +0000 UTC m=+2438.009331675" watchObservedRunningTime="2026-03-09 10:00:37.190215453 +0000 UTC m=+2438.014951034" Mar 09 10:00:38 crc kubenswrapper[4692]: I0309 10:00:38.181130 4692 generic.go:334] "Generic (PLEG): container finished" podID="2123c2c0-3966-4550-a5c1-9ece6e4d1e67" containerID="d5124acfaebcf56097f649b0f4498f5d7cccf9b3b630932cef47b05fa31c840a" exitCode=0 Mar 09 10:00:38 crc kubenswrapper[4692]: I0309 10:00:38.181211 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" event={"ID":"2123c2c0-3966-4550-a5c1-9ece6e4d1e67","Type":"ContainerDied","Data":"d5124acfaebcf56097f649b0f4498f5d7cccf9b3b630932cef47b05fa31c840a"} Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.479367 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.519215 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt"] Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.526295 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt"] Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.579218 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-etc-swift\") pod \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.579330 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-swiftconf\") pod \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.579373 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-ring-data-devices\") pod \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.579483 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl8rw\" (UniqueName: \"kubernetes.io/projected/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-kube-api-access-wl8rw\") pod \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.579613 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-scripts\") pod \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.579697 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-dispersionconf\") pod \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\" (UID: \"2123c2c0-3966-4550-a5c1-9ece6e4d1e67\") " Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.581409 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2123c2c0-3966-4550-a5c1-9ece6e4d1e67" (UID: "2123c2c0-3966-4550-a5c1-9ece6e4d1e67"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.581739 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2123c2c0-3966-4550-a5c1-9ece6e4d1e67" (UID: "2123c2c0-3966-4550-a5c1-9ece6e4d1e67"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.586153 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-kube-api-access-wl8rw" (OuterVolumeSpecName: "kube-api-access-wl8rw") pod "2123c2c0-3966-4550-a5c1-9ece6e4d1e67" (UID: "2123c2c0-3966-4550-a5c1-9ece6e4d1e67"). InnerVolumeSpecName "kube-api-access-wl8rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.602647 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-scripts" (OuterVolumeSpecName: "scripts") pod "2123c2c0-3966-4550-a5c1-9ece6e4d1e67" (UID: "2123c2c0-3966-4550-a5c1-9ece6e4d1e67"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.603760 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2123c2c0-3966-4550-a5c1-9ece6e4d1e67" (UID: "2123c2c0-3966-4550-a5c1-9ece6e4d1e67"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.605154 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2123c2c0-3966-4550-a5c1-9ece6e4d1e67" (UID: "2123c2c0-3966-4550-a5c1-9ece6e4d1e67"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.682265 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.682306 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.682319 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.682328 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.682337 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:39 crc kubenswrapper[4692]: I0309 10:00:39.682349 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl8rw\" (UniqueName: \"kubernetes.io/projected/2123c2c0-3966-4550-a5c1-9ece6e4d1e67-kube-api-access-wl8rw\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.082008 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2123c2c0-3966-4550-a5c1-9ece6e4d1e67" path="/var/lib/kubelet/pods/2123c2c0-3966-4550-a5c1-9ece6e4d1e67/volumes" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.203278 4692 scope.go:117] "RemoveContainer" containerID="d5124acfaebcf56097f649b0f4498f5d7cccf9b3b630932cef47b05fa31c840a" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.203336 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-wtzqt" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.688866 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2"] Mar 09 10:00:40 crc kubenswrapper[4692]: E0309 10:00:40.689706 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2123c2c0-3966-4550-a5c1-9ece6e4d1e67" containerName="swift-ring-rebalance" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.689722 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2123c2c0-3966-4550-a5c1-9ece6e4d1e67" containerName="swift-ring-rebalance" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.689861 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2123c2c0-3966-4550-a5c1-9ece6e4d1e67" containerName="swift-ring-rebalance" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.690465 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.695930 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.695979 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.704131 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2"] Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.804022 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kz8x\" (UniqueName: \"kubernetes.io/projected/96476cd2-32f4-4767-b06a-224fd227db4b-kube-api-access-8kz8x\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.804087 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-dispersionconf\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.804122 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96476cd2-32f4-4767-b06a-224fd227db4b-etc-swift\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.804214 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.804264 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-swiftconf\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.804345 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-scripts\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.906355 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-scripts\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.906734 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kz8x\" (UniqueName: \"kubernetes.io/projected/96476cd2-32f4-4767-b06a-224fd227db4b-kube-api-access-8kz8x\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.906826 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-dispersionconf\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.906909 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96476cd2-32f4-4767-b06a-224fd227db4b-etc-swift\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.906997 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.907088 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-swiftconf\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.907111 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-scripts\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.907505 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96476cd2-32f4-4767-b06a-224fd227db4b-etc-swift\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.907659 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.915043 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-swiftconf\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.922208 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-dispersionconf\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:40 crc kubenswrapper[4692]: I0309 10:00:40.927604 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kz8x\" (UniqueName: \"kubernetes.io/projected/96476cd2-32f4-4767-b06a-224fd227db4b-kube-api-access-8kz8x\") pod \"swift-ring-rebalance-debug-hvkw2\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:41 crc kubenswrapper[4692]: I0309 10:00:41.014224 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:41 crc kubenswrapper[4692]: I0309 10:00:41.448408 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2"] Mar 09 10:00:41 crc kubenswrapper[4692]: W0309 10:00:41.458801 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96476cd2_32f4_4767_b06a_224fd227db4b.slice/crio-a2f382a894fbc9edbfbfaed059c3fa63f14247ccea329e8fe8e2f6e7b464e2db WatchSource:0}: Error finding container a2f382a894fbc9edbfbfaed059c3fa63f14247ccea329e8fe8e2f6e7b464e2db: Status 404 returned error can't find the container with id a2f382a894fbc9edbfbfaed059c3fa63f14247ccea329e8fe8e2f6e7b464e2db Mar 09 10:00:42 crc kubenswrapper[4692]: I0309 10:00:42.072023 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:00:42 crc kubenswrapper[4692]: E0309 10:00:42.072743 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:00:42 crc kubenswrapper[4692]: I0309 10:00:42.229546 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" event={"ID":"96476cd2-32f4-4767-b06a-224fd227db4b","Type":"ContainerStarted","Data":"419051724af30ea3bdc1e1de61ef54d2dcc223abd644181bed2e505e86edf69f"} Mar 09 10:00:42 crc kubenswrapper[4692]: I0309 10:00:42.229604 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" event={"ID":"96476cd2-32f4-4767-b06a-224fd227db4b","Type":"ContainerStarted","Data":"a2f382a894fbc9edbfbfaed059c3fa63f14247ccea329e8fe8e2f6e7b464e2db"} Mar 09 10:00:43 crc kubenswrapper[4692]: I0309 10:00:43.240309 4692 generic.go:334] "Generic (PLEG): container finished" podID="96476cd2-32f4-4767-b06a-224fd227db4b" containerID="419051724af30ea3bdc1e1de61ef54d2dcc223abd644181bed2e505e86edf69f" exitCode=0 Mar 09 10:00:43 crc kubenswrapper[4692]: I0309 10:00:43.240396 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" event={"ID":"96476cd2-32f4-4767-b06a-224fd227db4b","Type":"ContainerDied","Data":"419051724af30ea3bdc1e1de61ef54d2dcc223abd644181bed2e505e86edf69f"} Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.516363 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.566207 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2"] Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.568466 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-dispersionconf\") pod \"96476cd2-32f4-4767-b06a-224fd227db4b\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.568527 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96476cd2-32f4-4767-b06a-224fd227db4b-etc-swift\") pod \"96476cd2-32f4-4767-b06a-224fd227db4b\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.568587 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-scripts\") pod \"96476cd2-32f4-4767-b06a-224fd227db4b\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.568704 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-swiftconf\") pod \"96476cd2-32f4-4767-b06a-224fd227db4b\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.568793 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kz8x\" (UniqueName: \"kubernetes.io/projected/96476cd2-32f4-4767-b06a-224fd227db4b-kube-api-access-8kz8x\") pod \"96476cd2-32f4-4767-b06a-224fd227db4b\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.568832 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-ring-data-devices\") pod \"96476cd2-32f4-4767-b06a-224fd227db4b\" (UID: \"96476cd2-32f4-4767-b06a-224fd227db4b\") " Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.569752 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "96476cd2-32f4-4767-b06a-224fd227db4b" (UID: "96476cd2-32f4-4767-b06a-224fd227db4b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.569981 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96476cd2-32f4-4767-b06a-224fd227db4b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "96476cd2-32f4-4767-b06a-224fd227db4b" (UID: "96476cd2-32f4-4767-b06a-224fd227db4b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.574596 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2"] Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.576325 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96476cd2-32f4-4767-b06a-224fd227db4b-kube-api-access-8kz8x" (OuterVolumeSpecName: "kube-api-access-8kz8x") pod "96476cd2-32f4-4767-b06a-224fd227db4b" (UID: "96476cd2-32f4-4767-b06a-224fd227db4b"). InnerVolumeSpecName "kube-api-access-8kz8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.595983 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "96476cd2-32f4-4767-b06a-224fd227db4b" (UID: "96476cd2-32f4-4767-b06a-224fd227db4b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.596186 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "96476cd2-32f4-4767-b06a-224fd227db4b" (UID: "96476cd2-32f4-4767-b06a-224fd227db4b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.599619 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-scripts" (OuterVolumeSpecName: "scripts") pod "96476cd2-32f4-4767-b06a-224fd227db4b" (UID: "96476cd2-32f4-4767-b06a-224fd227db4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.670852 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.670896 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/96476cd2-32f4-4767-b06a-224fd227db4b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.670906 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.670916 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/96476cd2-32f4-4767-b06a-224fd227db4b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.670927 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kz8x\" (UniqueName: \"kubernetes.io/projected/96476cd2-32f4-4767-b06a-224fd227db4b-kube-api-access-8kz8x\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:44 crc kubenswrapper[4692]: I0309 10:00:44.670938 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/96476cd2-32f4-4767-b06a-224fd227db4b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.262553 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2f382a894fbc9edbfbfaed059c3fa63f14247ccea329e8fe8e2f6e7b464e2db" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.262711 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-hvkw2" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.738896 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-94qp5"] Mar 09 10:00:45 crc kubenswrapper[4692]: E0309 10:00:45.739254 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96476cd2-32f4-4767-b06a-224fd227db4b" containerName="swift-ring-rebalance" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.739268 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="96476cd2-32f4-4767-b06a-224fd227db4b" containerName="swift-ring-rebalance" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.739428 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="96476cd2-32f4-4767-b06a-224fd227db4b" containerName="swift-ring-rebalance" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.740014 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.742180 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.742923 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.757784 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-94qp5"] Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.789622 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-dispersionconf\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.789692 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-ring-data-devices\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.789798 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56p9h\" (UniqueName: \"kubernetes.io/projected/26825c7d-090e-4fd7-a988-cbd147338ffd-kube-api-access-56p9h\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.789835 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-swiftconf\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.789870 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-scripts\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.789889 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26825c7d-090e-4fd7-a988-cbd147338ffd-etc-swift\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.891798 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-ring-data-devices\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.891923 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56p9h\" (UniqueName: \"kubernetes.io/projected/26825c7d-090e-4fd7-a988-cbd147338ffd-kube-api-access-56p9h\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.891966 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-swiftconf\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.891999 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-scripts\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.892018 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26825c7d-090e-4fd7-a988-cbd147338ffd-etc-swift\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.892053 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-dispersionconf\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.892918 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26825c7d-090e-4fd7-a988-cbd147338ffd-etc-swift\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.892992 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-ring-data-devices\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.893125 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-scripts\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.896720 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-dispersionconf\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.908687 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-swiftconf\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:45 crc kubenswrapper[4692]: I0309 10:00:45.912216 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56p9h\" (UniqueName: \"kubernetes.io/projected/26825c7d-090e-4fd7-a988-cbd147338ffd-kube-api-access-56p9h\") pod \"swift-ring-rebalance-debug-94qp5\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:46 crc kubenswrapper[4692]: I0309 10:00:46.065136 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:46 crc kubenswrapper[4692]: I0309 10:00:46.083708 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96476cd2-32f4-4767-b06a-224fd227db4b" path="/var/lib/kubelet/pods/96476cd2-32f4-4767-b06a-224fd227db4b/volumes" Mar 09 10:00:46 crc kubenswrapper[4692]: I0309 10:00:46.524503 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-94qp5"] Mar 09 10:00:46 crc kubenswrapper[4692]: W0309 10:00:46.531468 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26825c7d_090e_4fd7_a988_cbd147338ffd.slice/crio-71424e65159819c9337b05c8a3ab4ce55aa77230a6f9d5fb97a7065b774d2e7d WatchSource:0}: Error finding container 71424e65159819c9337b05c8a3ab4ce55aa77230a6f9d5fb97a7065b774d2e7d: Status 404 returned error can't find the container with id 71424e65159819c9337b05c8a3ab4ce55aa77230a6f9d5fb97a7065b774d2e7d Mar 09 10:00:47 crc kubenswrapper[4692]: I0309 10:00:47.284115 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" event={"ID":"26825c7d-090e-4fd7-a988-cbd147338ffd","Type":"ContainerStarted","Data":"50c1f2a768a8c75c0b3ae6c055543d888ed772413ddeddf77580046337d4b322"} Mar 09 10:00:47 crc kubenswrapper[4692]: I0309 10:00:47.284187 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" event={"ID":"26825c7d-090e-4fd7-a988-cbd147338ffd","Type":"ContainerStarted","Data":"71424e65159819c9337b05c8a3ab4ce55aa77230a6f9d5fb97a7065b774d2e7d"} Mar 09 10:00:47 crc kubenswrapper[4692]: I0309 10:00:47.306771 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" podStartSLOduration=2.306741335 podStartE2EDuration="2.306741335s" podCreationTimestamp="2026-03-09 10:00:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:47.305035736 +0000 UTC m=+2448.129771367" watchObservedRunningTime="2026-03-09 10:00:47.306741335 +0000 UTC m=+2448.131476936" Mar 09 10:00:48 crc kubenswrapper[4692]: I0309 10:00:48.295026 4692 generic.go:334] "Generic (PLEG): container finished" podID="26825c7d-090e-4fd7-a988-cbd147338ffd" containerID="50c1f2a768a8c75c0b3ae6c055543d888ed772413ddeddf77580046337d4b322" exitCode=0 Mar 09 10:00:48 crc kubenswrapper[4692]: I0309 10:00:48.295137 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" event={"ID":"26825c7d-090e-4fd7-a988-cbd147338ffd","Type":"ContainerDied","Data":"50c1f2a768a8c75c0b3ae6c055543d888ed772413ddeddf77580046337d4b322"} Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.586185 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.622496 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-94qp5"] Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.632615 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-94qp5"] Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.655458 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56p9h\" (UniqueName: \"kubernetes.io/projected/26825c7d-090e-4fd7-a988-cbd147338ffd-kube-api-access-56p9h\") pod \"26825c7d-090e-4fd7-a988-cbd147338ffd\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.655571 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-scripts\") pod \"26825c7d-090e-4fd7-a988-cbd147338ffd\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.655652 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-swiftconf\") pod \"26825c7d-090e-4fd7-a988-cbd147338ffd\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.655691 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-ring-data-devices\") pod \"26825c7d-090e-4fd7-a988-cbd147338ffd\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.655739 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-dispersionconf\") pod \"26825c7d-090e-4fd7-a988-cbd147338ffd\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.655794 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26825c7d-090e-4fd7-a988-cbd147338ffd-etc-swift\") pod \"26825c7d-090e-4fd7-a988-cbd147338ffd\" (UID: \"26825c7d-090e-4fd7-a988-cbd147338ffd\") " Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.656998 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26825c7d-090e-4fd7-a988-cbd147338ffd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "26825c7d-090e-4fd7-a988-cbd147338ffd" (UID: "26825c7d-090e-4fd7-a988-cbd147338ffd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.657807 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "26825c7d-090e-4fd7-a988-cbd147338ffd" (UID: "26825c7d-090e-4fd7-a988-cbd147338ffd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.664379 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26825c7d-090e-4fd7-a988-cbd147338ffd-kube-api-access-56p9h" (OuterVolumeSpecName: "kube-api-access-56p9h") pod "26825c7d-090e-4fd7-a988-cbd147338ffd" (UID: "26825c7d-090e-4fd7-a988-cbd147338ffd"). InnerVolumeSpecName "kube-api-access-56p9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.680205 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-scripts" (OuterVolumeSpecName: "scripts") pod "26825c7d-090e-4fd7-a988-cbd147338ffd" (UID: "26825c7d-090e-4fd7-a988-cbd147338ffd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.689338 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "26825c7d-090e-4fd7-a988-cbd147338ffd" (UID: "26825c7d-090e-4fd7-a988-cbd147338ffd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.691665 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "26825c7d-090e-4fd7-a988-cbd147338ffd" (UID: "26825c7d-090e-4fd7-a988-cbd147338ffd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.757299 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56p9h\" (UniqueName: \"kubernetes.io/projected/26825c7d-090e-4fd7-a988-cbd147338ffd-kube-api-access-56p9h\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.757348 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.757357 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.757370 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/26825c7d-090e-4fd7-a988-cbd147338ffd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.757379 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/26825c7d-090e-4fd7-a988-cbd147338ffd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:49 crc kubenswrapper[4692]: I0309 10:00:49.757387 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/26825c7d-090e-4fd7-a988-cbd147338ffd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.080737 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26825c7d-090e-4fd7-a988-cbd147338ffd" path="/var/lib/kubelet/pods/26825c7d-090e-4fd7-a988-cbd147338ffd/volumes" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.319484 4692 scope.go:117] "RemoveContainer" containerID="50c1f2a768a8c75c0b3ae6c055543d888ed772413ddeddf77580046337d4b322" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.319664 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-94qp5" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.802212 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv"] Mar 09 10:00:50 crc kubenswrapper[4692]: E0309 10:00:50.803590 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26825c7d-090e-4fd7-a988-cbd147338ffd" containerName="swift-ring-rebalance" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.803708 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="26825c7d-090e-4fd7-a988-cbd147338ffd" containerName="swift-ring-rebalance" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.803913 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="26825c7d-090e-4fd7-a988-cbd147338ffd" containerName="swift-ring-rebalance" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.804537 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.807424 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.807903 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.815656 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv"] Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.874805 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-ring-data-devices\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.874878 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/707968c5-c1f6-4da1-a988-dd8cf830c617-etc-swift\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.874900 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-scripts\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.874931 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m25g\" (UniqueName: \"kubernetes.io/projected/707968c5-c1f6-4da1-a988-dd8cf830c617-kube-api-access-8m25g\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.874996 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-dispersionconf\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.875506 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-swiftconf\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.976700 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-swiftconf\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.976776 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-ring-data-devices\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.976810 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/707968c5-c1f6-4da1-a988-dd8cf830c617-etc-swift\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.976827 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-scripts\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.976854 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m25g\" (UniqueName: \"kubernetes.io/projected/707968c5-c1f6-4da1-a988-dd8cf830c617-kube-api-access-8m25g\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.976876 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-dispersionconf\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.977840 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/707968c5-c1f6-4da1-a988-dd8cf830c617-etc-swift\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.978366 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-ring-data-devices\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.978375 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-scripts\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.981245 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-dispersionconf\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.981522 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-swiftconf\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:50 crc kubenswrapper[4692]: I0309 10:00:50.995332 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m25g\" (UniqueName: \"kubernetes.io/projected/707968c5-c1f6-4da1-a988-dd8cf830c617-kube-api-access-8m25g\") pod \"swift-ring-rebalance-debug-8zcnv\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:51 crc kubenswrapper[4692]: I0309 10:00:51.133886 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:51 crc kubenswrapper[4692]: I0309 10:00:51.590990 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv"] Mar 09 10:00:52 crc kubenswrapper[4692]: I0309 10:00:52.347842 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" event={"ID":"707968c5-c1f6-4da1-a988-dd8cf830c617","Type":"ContainerStarted","Data":"1b197625f7ebd47c1a1c9b1d61e82bcda624a5305d0f6315a1ee4595981726aa"} Mar 09 10:00:52 crc kubenswrapper[4692]: I0309 10:00:52.347919 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" event={"ID":"707968c5-c1f6-4da1-a988-dd8cf830c617","Type":"ContainerStarted","Data":"b2275ec4ef2149fa75adb5b30b53743a06a7af64a3b4aead162182337725807a"} Mar 09 10:00:52 crc kubenswrapper[4692]: I0309 10:00:52.374357 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" podStartSLOduration=2.374326609 podStartE2EDuration="2.374326609s" podCreationTimestamp="2026-03-09 10:00:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:52.369120212 +0000 UTC m=+2453.193855803" watchObservedRunningTime="2026-03-09 10:00:52.374326609 +0000 UTC m=+2453.199062190" Mar 09 10:00:53 crc kubenswrapper[4692]: I0309 10:00:53.360396 4692 generic.go:334] "Generic (PLEG): container finished" podID="707968c5-c1f6-4da1-a988-dd8cf830c617" containerID="1b197625f7ebd47c1a1c9b1d61e82bcda624a5305d0f6315a1ee4595981726aa" exitCode=0 Mar 09 10:00:53 crc kubenswrapper[4692]: I0309 10:00:53.360554 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" event={"ID":"707968c5-c1f6-4da1-a988-dd8cf830c617","Type":"ContainerDied","Data":"1b197625f7ebd47c1a1c9b1d61e82bcda624a5305d0f6315a1ee4595981726aa"} Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.699127 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.796726 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv"] Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.802932 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv"] Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.891648 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-dispersionconf\") pod \"707968c5-c1f6-4da1-a988-dd8cf830c617\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.891758 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-swiftconf\") pod \"707968c5-c1f6-4da1-a988-dd8cf830c617\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.891822 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-scripts\") pod \"707968c5-c1f6-4da1-a988-dd8cf830c617\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.891925 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-ring-data-devices\") pod \"707968c5-c1f6-4da1-a988-dd8cf830c617\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.891968 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/707968c5-c1f6-4da1-a988-dd8cf830c617-etc-swift\") pod \"707968c5-c1f6-4da1-a988-dd8cf830c617\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.892047 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m25g\" (UniqueName: \"kubernetes.io/projected/707968c5-c1f6-4da1-a988-dd8cf830c617-kube-api-access-8m25g\") pod \"707968c5-c1f6-4da1-a988-dd8cf830c617\" (UID: \"707968c5-c1f6-4da1-a988-dd8cf830c617\") " Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.893011 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/707968c5-c1f6-4da1-a988-dd8cf830c617-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "707968c5-c1f6-4da1-a988-dd8cf830c617" (UID: "707968c5-c1f6-4da1-a988-dd8cf830c617"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.894434 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "707968c5-c1f6-4da1-a988-dd8cf830c617" (UID: "707968c5-c1f6-4da1-a988-dd8cf830c617"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.900475 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/707968c5-c1f6-4da1-a988-dd8cf830c617-kube-api-access-8m25g" (OuterVolumeSpecName: "kube-api-access-8m25g") pod "707968c5-c1f6-4da1-a988-dd8cf830c617" (UID: "707968c5-c1f6-4da1-a988-dd8cf830c617"). InnerVolumeSpecName "kube-api-access-8m25g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.915431 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-scripts" (OuterVolumeSpecName: "scripts") pod "707968c5-c1f6-4da1-a988-dd8cf830c617" (UID: "707968c5-c1f6-4da1-a988-dd8cf830c617"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.918085 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "707968c5-c1f6-4da1-a988-dd8cf830c617" (UID: "707968c5-c1f6-4da1-a988-dd8cf830c617"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.923170 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "707968c5-c1f6-4da1-a988-dd8cf830c617" (UID: "707968c5-c1f6-4da1-a988-dd8cf830c617"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.994581 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.994631 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/707968c5-c1f6-4da1-a988-dd8cf830c617-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.994647 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/707968c5-c1f6-4da1-a988-dd8cf830c617-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.994658 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m25g\" (UniqueName: \"kubernetes.io/projected/707968c5-c1f6-4da1-a988-dd8cf830c617-kube-api-access-8m25g\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.994670 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:54 crc kubenswrapper[4692]: I0309 10:00:54.994680 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/707968c5-c1f6-4da1-a988-dd8cf830c617-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.071407 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:00:55 crc kubenswrapper[4692]: E0309 10:00:55.071671 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.388420 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2275ec4ef2149fa75adb5b30b53743a06a7af64a3b4aead162182337725807a" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.388572 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-8zcnv" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.934047 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2"] Mar 09 10:00:55 crc kubenswrapper[4692]: E0309 10:00:55.935628 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="707968c5-c1f6-4da1-a988-dd8cf830c617" containerName="swift-ring-rebalance" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.935733 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="707968c5-c1f6-4da1-a988-dd8cf830c617" containerName="swift-ring-rebalance" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.936099 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="707968c5-c1f6-4da1-a988-dd8cf830c617" containerName="swift-ring-rebalance" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.936945 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.941588 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.942296 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:00:55 crc kubenswrapper[4692]: I0309 10:00:55.946404 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2"] Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.081416 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="707968c5-c1f6-4da1-a988-dd8cf830c617" path="/var/lib/kubelet/pods/707968c5-c1f6-4da1-a988-dd8cf830c617/volumes" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.117971 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-dispersionconf\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.118026 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2dl5\" (UniqueName: \"kubernetes.io/projected/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-kube-api-access-n2dl5\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.118060 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-etc-swift\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.118097 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-scripts\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.118115 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-swiftconf\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.118148 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-ring-data-devices\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.219918 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-dispersionconf\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.219976 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2dl5\" (UniqueName: \"kubernetes.io/projected/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-kube-api-access-n2dl5\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.220003 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-etc-swift\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.220050 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-scripts\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.220072 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-swiftconf\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.220128 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-ring-data-devices\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.221147 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-ring-data-devices\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.223109 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-scripts\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.223410 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-etc-swift\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.228848 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-dispersionconf\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.229978 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-swiftconf\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.246180 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2dl5\" (UniqueName: \"kubernetes.io/projected/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-kube-api-access-n2dl5\") pod \"swift-ring-rebalance-debug-4kcp2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.255998 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:56 crc kubenswrapper[4692]: I0309 10:00:56.507925 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2"] Mar 09 10:00:56 crc kubenswrapper[4692]: W0309 10:00:56.515630 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d86bcbf_f2c1_4253_b3bc_0b76d6bb23a2.slice/crio-da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f WatchSource:0}: Error finding container da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f: Status 404 returned error can't find the container with id da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f Mar 09 10:00:57 crc kubenswrapper[4692]: I0309 10:00:57.413206 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" event={"ID":"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2","Type":"ContainerStarted","Data":"daf27c055f165ce3eb7e2fdf5c8c6562ccafa2493663ce4def0ee6639e744114"} Mar 09 10:00:57 crc kubenswrapper[4692]: I0309 10:00:57.413295 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" event={"ID":"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2","Type":"ContainerStarted","Data":"da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f"} Mar 09 10:00:57 crc kubenswrapper[4692]: I0309 10:00:57.434754 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" podStartSLOduration=2.434728691 podStartE2EDuration="2.434728691s" podCreationTimestamp="2026-03-09 10:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:00:57.430039729 +0000 UTC m=+2458.254775310" watchObservedRunningTime="2026-03-09 10:00:57.434728691 +0000 UTC m=+2458.259464272" Mar 09 10:00:58 crc kubenswrapper[4692]: I0309 10:00:58.426370 4692 generic.go:334] "Generic (PLEG): container finished" podID="6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" containerID="daf27c055f165ce3eb7e2fdf5c8c6562ccafa2493663ce4def0ee6639e744114" exitCode=0 Mar 09 10:00:58 crc kubenswrapper[4692]: I0309 10:00:58.426437 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" event={"ID":"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2","Type":"ContainerDied","Data":"daf27c055f165ce3eb7e2fdf5c8c6562ccafa2493663ce4def0ee6639e744114"} Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.724102 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.763209 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2"] Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.779888 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2"] Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.887752 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2dl5\" (UniqueName: \"kubernetes.io/projected/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-kube-api-access-n2dl5\") pod \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.887839 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-dispersionconf\") pod \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.888000 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-etc-swift\") pod \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.888025 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-ring-data-devices\") pod \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.888133 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-scripts\") pod \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.888563 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" (UID: "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.888609 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-swiftconf\") pod \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\" (UID: \"6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2\") " Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.888802 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" (UID: "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.889534 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.889566 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.895211 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-kube-api-access-n2dl5" (OuterVolumeSpecName: "kube-api-access-n2dl5") pod "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" (UID: "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2"). InnerVolumeSpecName "kube-api-access-n2dl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.910761 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-scripts" (OuterVolumeSpecName: "scripts") pod "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" (UID: "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.914038 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" (UID: "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.923346 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" (UID: "6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.991694 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.991738 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.991751 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2dl5\" (UniqueName: \"kubernetes.io/projected/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-kube-api-access-n2dl5\") on node \"crc\" DevicePath \"\"" Mar 09 10:00:59 crc kubenswrapper[4692]: I0309 10:00:59.991764 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.084364 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" path="/var/lib/kubelet/pods/6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2/volumes" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.149485 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/keystone-cron-29550841-psdz4"] Mar 09 10:01:00 crc kubenswrapper[4692]: E0309 10:01:00.150044 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" containerName="swift-ring-rebalance" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.150096 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" containerName="swift-ring-rebalance" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.150337 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d86bcbf-f2c1-4253-b3bc-0b76d6bb23a2" containerName="swift-ring-rebalance" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.151097 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.164054 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-cron-29550841-psdz4"] Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.297766 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xn9x\" (UniqueName: \"kubernetes.io/projected/e4861760-6b7f-4777-a35d-bdb3614c4d39-kube-api-access-9xn9x\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.297829 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-fernet-keys\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.298765 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-config-data\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.400371 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xn9x\" (UniqueName: \"kubernetes.io/projected/e4861760-6b7f-4777-a35d-bdb3614c4d39-kube-api-access-9xn9x\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.400440 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-fernet-keys\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.400519 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-config-data\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.406835 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-fernet-keys\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.407226 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-config-data\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.418557 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xn9x\" (UniqueName: \"kubernetes.io/projected/e4861760-6b7f-4777-a35d-bdb3614c4d39-kube-api-access-9xn9x\") pod \"keystone-cron-29550841-psdz4\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.447051 4692 scope.go:117] "RemoveContainer" containerID="daf27c055f165ce3eb7e2fdf5c8c6562ccafa2493663ce4def0ee6639e744114" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.447295 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4kcp2" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.473831 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.855260 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/keystone-cron-29550841-psdz4"] Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.940622 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr"] Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.942373 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.944979 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.949429 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:00 crc kubenswrapper[4692]: I0309 10:01:00.955640 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr"] Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.013859 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ee561fa-9c4c-4390-b7c3-068c166705d0-etc-swift\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.013943 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lnp9\" (UniqueName: \"kubernetes.io/projected/3ee561fa-9c4c-4390-b7c3-068c166705d0-kube-api-access-8lnp9\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.013987 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-ring-data-devices\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.014013 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-dispersionconf\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.014074 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-scripts\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.014361 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-swiftconf\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.116263 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ee561fa-9c4c-4390-b7c3-068c166705d0-etc-swift\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.116380 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lnp9\" (UniqueName: \"kubernetes.io/projected/3ee561fa-9c4c-4390-b7c3-068c166705d0-kube-api-access-8lnp9\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.116416 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-ring-data-devices\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.116450 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-dispersionconf\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.116536 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-scripts\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.116561 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-swiftconf\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.117671 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-scripts\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.117716 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-ring-data-devices\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.117938 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ee561fa-9c4c-4390-b7c3-068c166705d0-etc-swift\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.121816 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-dispersionconf\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.123205 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-swiftconf\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.136550 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lnp9\" (UniqueName: \"kubernetes.io/projected/3ee561fa-9c4c-4390-b7c3-068c166705d0-kube-api-access-8lnp9\") pod \"swift-ring-rebalance-debug-kbcvr\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.305961 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.461887 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" event={"ID":"e4861760-6b7f-4777-a35d-bdb3614c4d39","Type":"ContainerStarted","Data":"1c13ad2cf7eeb58328b3b8bab0850407a62da9e93796a4ddc29fa8aa042cf88d"} Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.461987 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" event={"ID":"e4861760-6b7f-4777-a35d-bdb3614c4d39","Type":"ContainerStarted","Data":"39777060a0200b94b13049ad440cd6aeb86f0d4a305fd8f9615e6a55b76a8b12"} Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.486307 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" podStartSLOduration=1.486279266 podStartE2EDuration="1.486279266s" podCreationTimestamp="2026-03-09 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:01.479872736 +0000 UTC m=+2462.304608317" watchObservedRunningTime="2026-03-09 10:01:01.486279266 +0000 UTC m=+2462.311014847" Mar 09 10:01:01 crc kubenswrapper[4692]: I0309 10:01:01.571028 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr"] Mar 09 10:01:01 crc kubenswrapper[4692]: W0309 10:01:01.577062 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ee561fa_9c4c_4390_b7c3_068c166705d0.slice/crio-2d93622944012c9ed24a1a8826470a07cd204bad6b5461541802086fea119eb7 WatchSource:0}: Error finding container 2d93622944012c9ed24a1a8826470a07cd204bad6b5461541802086fea119eb7: Status 404 returned error can't find the container with id 2d93622944012c9ed24a1a8826470a07cd204bad6b5461541802086fea119eb7 Mar 09 10:01:02 crc kubenswrapper[4692]: E0309 10:01:02.468363 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d86bcbf_f2c1_4253_b3bc_0b76d6bb23a2.slice/crio-da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f\": RecentStats: unable to find data in memory cache]" Mar 09 10:01:02 crc kubenswrapper[4692]: I0309 10:01:02.483003 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" event={"ID":"3ee561fa-9c4c-4390-b7c3-068c166705d0","Type":"ContainerStarted","Data":"f369fed400d69274d22d79a1e9711d75aac725496a6acf9315c98aed71d2508c"} Mar 09 10:01:02 crc kubenswrapper[4692]: I0309 10:01:02.483072 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" event={"ID":"3ee561fa-9c4c-4390-b7c3-068c166705d0","Type":"ContainerStarted","Data":"2d93622944012c9ed24a1a8826470a07cd204bad6b5461541802086fea119eb7"} Mar 09 10:01:02 crc kubenswrapper[4692]: I0309 10:01:02.520948 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" podStartSLOduration=2.52092086 podStartE2EDuration="2.52092086s" podCreationTimestamp="2026-03-09 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:02.510571199 +0000 UTC m=+2463.335306780" watchObservedRunningTime="2026-03-09 10:01:02.52092086 +0000 UTC m=+2463.345656441" Mar 09 10:01:03 crc kubenswrapper[4692]: I0309 10:01:03.494698 4692 generic.go:334] "Generic (PLEG): container finished" podID="3ee561fa-9c4c-4390-b7c3-068c166705d0" containerID="f369fed400d69274d22d79a1e9711d75aac725496a6acf9315c98aed71d2508c" exitCode=0 Mar 09 10:01:03 crc kubenswrapper[4692]: I0309 10:01:03.494804 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" event={"ID":"3ee561fa-9c4c-4390-b7c3-068c166705d0","Type":"ContainerDied","Data":"f369fed400d69274d22d79a1e9711d75aac725496a6acf9315c98aed71d2508c"} Mar 09 10:01:03 crc kubenswrapper[4692]: I0309 10:01:03.498036 4692 generic.go:334] "Generic (PLEG): container finished" podID="e4861760-6b7f-4777-a35d-bdb3614c4d39" containerID="1c13ad2cf7eeb58328b3b8bab0850407a62da9e93796a4ddc29fa8aa042cf88d" exitCode=0 Mar 09 10:01:03 crc kubenswrapper[4692]: I0309 10:01:03.498082 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" event={"ID":"e4861760-6b7f-4777-a35d-bdb3614c4d39","Type":"ContainerDied","Data":"1c13ad2cf7eeb58328b3b8bab0850407a62da9e93796a4ddc29fa8aa042cf88d"} Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.879255 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.894825 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.944054 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr"] Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.951583 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr"] Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.991228 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-config-data\") pod \"e4861760-6b7f-4777-a35d-bdb3614c4d39\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.991358 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-fernet-keys\") pod \"e4861760-6b7f-4777-a35d-bdb3614c4d39\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.991417 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xn9x\" (UniqueName: \"kubernetes.io/projected/e4861760-6b7f-4777-a35d-bdb3614c4d39-kube-api-access-9xn9x\") pod \"e4861760-6b7f-4777-a35d-bdb3614c4d39\" (UID: \"e4861760-6b7f-4777-a35d-bdb3614c4d39\") " Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.998114 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4861760-6b7f-4777-a35d-bdb3614c4d39-kube-api-access-9xn9x" (OuterVolumeSpecName: "kube-api-access-9xn9x") pod "e4861760-6b7f-4777-a35d-bdb3614c4d39" (UID: "e4861760-6b7f-4777-a35d-bdb3614c4d39"). InnerVolumeSpecName "kube-api-access-9xn9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:04 crc kubenswrapper[4692]: I0309 10:01:04.998127 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e4861760-6b7f-4777-a35d-bdb3614c4d39" (UID: "e4861760-6b7f-4777-a35d-bdb3614c4d39"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.052895 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-config-data" (OuterVolumeSpecName: "config-data") pod "e4861760-6b7f-4777-a35d-bdb3614c4d39" (UID: "e4861760-6b7f-4777-a35d-bdb3614c4d39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.093275 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-ring-data-devices\") pod \"3ee561fa-9c4c-4390-b7c3-068c166705d0\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.093418 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-swiftconf\") pod \"3ee561fa-9c4c-4390-b7c3-068c166705d0\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.093985 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-scripts\") pod \"3ee561fa-9c4c-4390-b7c3-068c166705d0\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094034 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ee561fa-9c4c-4390-b7c3-068c166705d0-etc-swift\") pod \"3ee561fa-9c4c-4390-b7c3-068c166705d0\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094141 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lnp9\" (UniqueName: \"kubernetes.io/projected/3ee561fa-9c4c-4390-b7c3-068c166705d0-kube-api-access-8lnp9\") pod \"3ee561fa-9c4c-4390-b7c3-068c166705d0\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094151 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3ee561fa-9c4c-4390-b7c3-068c166705d0" (UID: "3ee561fa-9c4c-4390-b7c3-068c166705d0"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094260 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-dispersionconf\") pod \"3ee561fa-9c4c-4390-b7c3-068c166705d0\" (UID: \"3ee561fa-9c4c-4390-b7c3-068c166705d0\") " Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094833 4692 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-config-data\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094852 4692 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e4861760-6b7f-4777-a35d-bdb3614c4d39-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094873 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xn9x\" (UniqueName: \"kubernetes.io/projected/e4861760-6b7f-4777-a35d-bdb3614c4d39-kube-api-access-9xn9x\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094885 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.094844 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ee561fa-9c4c-4390-b7c3-068c166705d0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3ee561fa-9c4c-4390-b7c3-068c166705d0" (UID: "3ee561fa-9c4c-4390-b7c3-068c166705d0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.098285 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ee561fa-9c4c-4390-b7c3-068c166705d0-kube-api-access-8lnp9" (OuterVolumeSpecName: "kube-api-access-8lnp9") pod "3ee561fa-9c4c-4390-b7c3-068c166705d0" (UID: "3ee561fa-9c4c-4390-b7c3-068c166705d0"). InnerVolumeSpecName "kube-api-access-8lnp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.117624 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-scripts" (OuterVolumeSpecName: "scripts") pod "3ee561fa-9c4c-4390-b7c3-068c166705d0" (UID: "3ee561fa-9c4c-4390-b7c3-068c166705d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.121100 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3ee561fa-9c4c-4390-b7c3-068c166705d0" (UID: "3ee561fa-9c4c-4390-b7c3-068c166705d0"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.121882 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3ee561fa-9c4c-4390-b7c3-068c166705d0" (UID: "3ee561fa-9c4c-4390-b7c3-068c166705d0"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.196070 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.196153 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ee561fa-9c4c-4390-b7c3-068c166705d0-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.196192 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ee561fa-9c4c-4390-b7c3-068c166705d0-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.196206 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lnp9\" (UniqueName: \"kubernetes.io/projected/3ee561fa-9c4c-4390-b7c3-068c166705d0-kube-api-access-8lnp9\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.196220 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ee561fa-9c4c-4390-b7c3-068c166705d0-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.536814 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbcvr" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.536813 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d93622944012c9ed24a1a8826470a07cd204bad6b5461541802086fea119eb7" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.540121 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" event={"ID":"e4861760-6b7f-4777-a35d-bdb3614c4d39","Type":"ContainerDied","Data":"39777060a0200b94b13049ad440cd6aeb86f0d4a305fd8f9615e6a55b76a8b12"} Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.540216 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/keystone-cron-29550841-psdz4" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.540233 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39777060a0200b94b13049ad440cd6aeb86f0d4a305fd8f9615e6a55b76a8b12" Mar 09 10:01:05 crc kubenswrapper[4692]: I0309 10:01:05.998884 4692 scope.go:117] "RemoveContainer" containerID="e99aa3985d9f9af6c359f6a6b82df125a0064c276634fa986fb9014f44f47afa" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.033361 4692 scope.go:117] "RemoveContainer" containerID="a0cd5a083aca529493f9715f1fc4bd83fd47ccbc981f17d396663bdabf8682f2" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.072524 4692 scope.go:117] "RemoveContainer" containerID="ac247f9bfc4074d231deb121d854bad9b6a80025df752a0138fad95d1a0dbfb0" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.083782 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ee561fa-9c4c-4390-b7c3-068c166705d0" path="/var/lib/kubelet/pods/3ee561fa-9c4c-4390-b7c3-068c166705d0/volumes" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.097040 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg"] Mar 09 10:01:06 crc kubenswrapper[4692]: E0309 10:01:06.097657 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4861760-6b7f-4777-a35d-bdb3614c4d39" containerName="keystone-cron" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.097697 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4861760-6b7f-4777-a35d-bdb3614c4d39" containerName="keystone-cron" Mar 09 10:01:06 crc kubenswrapper[4692]: E0309 10:01:06.097721 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ee561fa-9c4c-4390-b7c3-068c166705d0" containerName="swift-ring-rebalance" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.097734 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ee561fa-9c4c-4390-b7c3-068c166705d0" containerName="swift-ring-rebalance" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.097998 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4861760-6b7f-4777-a35d-bdb3614c4d39" containerName="keystone-cron" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.098033 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ee561fa-9c4c-4390-b7c3-068c166705d0" containerName="swift-ring-rebalance" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.098887 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.110983 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-ring-data-devices\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.111040 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-dispersionconf\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.111582 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-swiftconf\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.111636 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/284d8555-509b-4643-a2ae-6bb957a410e4-etc-swift\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.111664 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-scripts\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.111715 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94rx5\" (UniqueName: \"kubernetes.io/projected/284d8555-509b-4643-a2ae-6bb957a410e4-kube-api-access-94rx5\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.113486 4692 scope.go:117] "RemoveContainer" containerID="43f05d768a18ca113aa566c302d271df491fefed968e2ee4a9f5257cbb1029cb" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.115609 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg"] Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.116246 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.116567 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.172368 4692 scope.go:117] "RemoveContainer" containerID="b7875fa5116cc38b75a9fd03aa64eab7c165c848b2e7a772b63e69b7be9e95c8" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.203835 4692 scope.go:117] "RemoveContainer" containerID="0870e7d45a9beb4862343e1a3b02cf223c1aa8ab6a20e1625b1d78e3458bf5de" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.212793 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-ring-data-devices\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.212838 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-dispersionconf\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.212895 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-swiftconf\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.212922 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/284d8555-509b-4643-a2ae-6bb957a410e4-etc-swift\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.212939 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-scripts\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.212963 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94rx5\" (UniqueName: \"kubernetes.io/projected/284d8555-509b-4643-a2ae-6bb957a410e4-kube-api-access-94rx5\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.213793 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-ring-data-devices\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.213804 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/284d8555-509b-4643-a2ae-6bb957a410e4-etc-swift\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.213917 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-scripts\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.221592 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-swiftconf\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.221634 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-dispersionconf\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.235787 4692 scope.go:117] "RemoveContainer" containerID="9100ffdd6a1b366abe731cfd8b84c0588f3568266469b712d02fc54357e3a499" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.237310 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94rx5\" (UniqueName: \"kubernetes.io/projected/284d8555-509b-4643-a2ae-6bb957a410e4-kube-api-access-94rx5\") pod \"swift-ring-rebalance-debug-6ndbg\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.268700 4692 scope.go:117] "RemoveContainer" containerID="2bf1f642be46a1f3d093f0f8c44dcae82602cde0cc59a31ee9b2330b1a22f29b" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.297934 4692 scope.go:117] "RemoveContainer" containerID="0b60bfd273f7b234f723038ea0b4f37f5816c4ed4723ad914a7ce98f7a07678d" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.328277 4692 scope.go:117] "RemoveContainer" containerID="e5c2c85fde01a89a44f176830513b4333f8361e5afea337f991e5af9d8971c6e" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.354158 4692 scope.go:117] "RemoveContainer" containerID="828b3f9bb0c31f366f65b0271e5e0710357d7d5c49125d03befa7f534fd78afa" Mar 09 10:01:06 crc kubenswrapper[4692]: I0309 10:01:06.483409 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:07 crc kubenswrapper[4692]: I0309 10:01:07.024093 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg"] Mar 09 10:01:07 crc kubenswrapper[4692]: I0309 10:01:07.580898 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" event={"ID":"284d8555-509b-4643-a2ae-6bb957a410e4","Type":"ContainerStarted","Data":"5ab7049a990d6a34b2fd27408c6fe7c669aba8e49d9d36f8bf91cbbd1569d964"} Mar 09 10:01:07 crc kubenswrapper[4692]: I0309 10:01:07.581467 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" event={"ID":"284d8555-509b-4643-a2ae-6bb957a410e4","Type":"ContainerStarted","Data":"b10e169a06b7d88f3f4c8019d3132f55e4cec0f7bac85db08b8d8ab3e0b850a5"} Mar 09 10:01:07 crc kubenswrapper[4692]: I0309 10:01:07.603079 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" podStartSLOduration=1.603051424 podStartE2EDuration="1.603051424s" podCreationTimestamp="2026-03-09 10:01:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:07.59723383 +0000 UTC m=+2468.421969451" watchObservedRunningTime="2026-03-09 10:01:07.603051424 +0000 UTC m=+2468.427787025" Mar 09 10:01:08 crc kubenswrapper[4692]: I0309 10:01:08.072676 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:01:08 crc kubenswrapper[4692]: E0309 10:01:08.072956 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:01:09 crc kubenswrapper[4692]: I0309 10:01:09.603622 4692 generic.go:334] "Generic (PLEG): container finished" podID="284d8555-509b-4643-a2ae-6bb957a410e4" containerID="5ab7049a990d6a34b2fd27408c6fe7c669aba8e49d9d36f8bf91cbbd1569d964" exitCode=0 Mar 09 10:01:09 crc kubenswrapper[4692]: I0309 10:01:09.603703 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" event={"ID":"284d8555-509b-4643-a2ae-6bb957a410e4","Type":"ContainerDied","Data":"5ab7049a990d6a34b2fd27408c6fe7c669aba8e49d9d36f8bf91cbbd1569d964"} Mar 09 10:01:10 crc kubenswrapper[4692]: I0309 10:01:10.916447 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:10 crc kubenswrapper[4692]: I0309 10:01:10.958362 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg"] Mar 09 10:01:10 crc kubenswrapper[4692]: I0309 10:01:10.965985 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg"] Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.089762 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94rx5\" (UniqueName: \"kubernetes.io/projected/284d8555-509b-4643-a2ae-6bb957a410e4-kube-api-access-94rx5\") pod \"284d8555-509b-4643-a2ae-6bb957a410e4\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.089936 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-ring-data-devices\") pod \"284d8555-509b-4643-a2ae-6bb957a410e4\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.089979 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-dispersionconf\") pod \"284d8555-509b-4643-a2ae-6bb957a410e4\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.090027 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/284d8555-509b-4643-a2ae-6bb957a410e4-etc-swift\") pod \"284d8555-509b-4643-a2ae-6bb957a410e4\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.090066 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-swiftconf\") pod \"284d8555-509b-4643-a2ae-6bb957a410e4\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.090135 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-scripts\") pod \"284d8555-509b-4643-a2ae-6bb957a410e4\" (UID: \"284d8555-509b-4643-a2ae-6bb957a410e4\") " Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.090920 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "284d8555-509b-4643-a2ae-6bb957a410e4" (UID: "284d8555-509b-4643-a2ae-6bb957a410e4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.091547 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/284d8555-509b-4643-a2ae-6bb957a410e4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "284d8555-509b-4643-a2ae-6bb957a410e4" (UID: "284d8555-509b-4643-a2ae-6bb957a410e4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.097383 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/284d8555-509b-4643-a2ae-6bb957a410e4-kube-api-access-94rx5" (OuterVolumeSpecName: "kube-api-access-94rx5") pod "284d8555-509b-4643-a2ae-6bb957a410e4" (UID: "284d8555-509b-4643-a2ae-6bb957a410e4"). InnerVolumeSpecName "kube-api-access-94rx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.113061 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-scripts" (OuterVolumeSpecName: "scripts") pod "284d8555-509b-4643-a2ae-6bb957a410e4" (UID: "284d8555-509b-4643-a2ae-6bb957a410e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.116392 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "284d8555-509b-4643-a2ae-6bb957a410e4" (UID: "284d8555-509b-4643-a2ae-6bb957a410e4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.119419 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "284d8555-509b-4643-a2ae-6bb957a410e4" (UID: "284d8555-509b-4643-a2ae-6bb957a410e4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.192224 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.192971 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.193015 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/284d8555-509b-4643-a2ae-6bb957a410e4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.193031 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/284d8555-509b-4643-a2ae-6bb957a410e4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.193042 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/284d8555-509b-4643-a2ae-6bb957a410e4-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.193060 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94rx5\" (UniqueName: \"kubernetes.io/projected/284d8555-509b-4643-a2ae-6bb957a410e4-kube-api-access-94rx5\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.629105 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b10e169a06b7d88f3f4c8019d3132f55e4cec0f7bac85db08b8d8ab3e0b850a5" Mar 09 10:01:11 crc kubenswrapper[4692]: I0309 10:01:11.629238 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6ndbg" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.083330 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="284d8555-509b-4643-a2ae-6bb957a410e4" path="/var/lib/kubelet/pods/284d8555-509b-4643-a2ae-6bb957a410e4/volumes" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.129379 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6blv4"] Mar 09 10:01:12 crc kubenswrapper[4692]: E0309 10:01:12.129806 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284d8555-509b-4643-a2ae-6bb957a410e4" containerName="swift-ring-rebalance" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.129833 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="284d8555-509b-4643-a2ae-6bb957a410e4" containerName="swift-ring-rebalance" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.130069 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="284d8555-509b-4643-a2ae-6bb957a410e4" containerName="swift-ring-rebalance" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.130891 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.133774 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.134124 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.143868 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6blv4"] Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.210805 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-ring-data-devices\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.210988 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-dispersionconf\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.211116 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-swiftconf\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.211155 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58pn8\" (UniqueName: \"kubernetes.io/projected/e43bef3b-68f9-4e72-bf21-b89723c3efb6-kube-api-access-58pn8\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.211236 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e43bef3b-68f9-4e72-bf21-b89723c3efb6-etc-swift\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.211261 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-scripts\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.313846 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-ring-data-devices\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.314498 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-dispersionconf\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.314549 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-swiftconf\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.314578 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58pn8\" (UniqueName: \"kubernetes.io/projected/e43bef3b-68f9-4e72-bf21-b89723c3efb6-kube-api-access-58pn8\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.314621 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e43bef3b-68f9-4e72-bf21-b89723c3efb6-etc-swift\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.314652 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-scripts\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.314829 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-ring-data-devices\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.315261 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e43bef3b-68f9-4e72-bf21-b89723c3efb6-etc-swift\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.315717 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-scripts\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.323338 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-swiftconf\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.323448 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-dispersionconf\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.340580 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58pn8\" (UniqueName: \"kubernetes.io/projected/e43bef3b-68f9-4e72-bf21-b89723c3efb6-kube-api-access-58pn8\") pod \"swift-ring-rebalance-debug-6blv4\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.453781 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:12 crc kubenswrapper[4692]: E0309 10:01:12.727465 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d86bcbf_f2c1_4253_b3bc_0b76d6bb23a2.slice/crio-da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f\": RecentStats: unable to find data in memory cache]" Mar 09 10:01:12 crc kubenswrapper[4692]: I0309 10:01:12.904898 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6blv4"] Mar 09 10:01:12 crc kubenswrapper[4692]: W0309 10:01:12.909852 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode43bef3b_68f9_4e72_bf21_b89723c3efb6.slice/crio-8cbb9745ca21d99ce4717cef6cdc68dac7b94fa5e4de579068fd54af7582e50b WatchSource:0}: Error finding container 8cbb9745ca21d99ce4717cef6cdc68dac7b94fa5e4de579068fd54af7582e50b: Status 404 returned error can't find the container with id 8cbb9745ca21d99ce4717cef6cdc68dac7b94fa5e4de579068fd54af7582e50b Mar 09 10:01:13 crc kubenswrapper[4692]: I0309 10:01:13.649768 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" event={"ID":"e43bef3b-68f9-4e72-bf21-b89723c3efb6","Type":"ContainerStarted","Data":"dfeaf99e5565cdce727a23b58d8e19eecb34722c4a27e39bdacca69aa9b733b2"} Mar 09 10:01:13 crc kubenswrapper[4692]: I0309 10:01:13.650302 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" event={"ID":"e43bef3b-68f9-4e72-bf21-b89723c3efb6","Type":"ContainerStarted","Data":"8cbb9745ca21d99ce4717cef6cdc68dac7b94fa5e4de579068fd54af7582e50b"} Mar 09 10:01:13 crc kubenswrapper[4692]: I0309 10:01:13.675428 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" podStartSLOduration=1.67540353 podStartE2EDuration="1.67540353s" podCreationTimestamp="2026-03-09 10:01:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:13.668094395 +0000 UTC m=+2474.492830006" watchObservedRunningTime="2026-03-09 10:01:13.67540353 +0000 UTC m=+2474.500139111" Mar 09 10:01:15 crc kubenswrapper[4692]: I0309 10:01:15.673103 4692 generic.go:334] "Generic (PLEG): container finished" podID="e43bef3b-68f9-4e72-bf21-b89723c3efb6" containerID="dfeaf99e5565cdce727a23b58d8e19eecb34722c4a27e39bdacca69aa9b733b2" exitCode=0 Mar 09 10:01:15 crc kubenswrapper[4692]: I0309 10:01:15.673211 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" event={"ID":"e43bef3b-68f9-4e72-bf21-b89723c3efb6","Type":"ContainerDied","Data":"dfeaf99e5565cdce727a23b58d8e19eecb34722c4a27e39bdacca69aa9b733b2"} Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.000675 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.050245 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6blv4"] Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.059480 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-6blv4"] Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.194371 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-swiftconf\") pod \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.194460 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-scripts\") pod \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.194490 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58pn8\" (UniqueName: \"kubernetes.io/projected/e43bef3b-68f9-4e72-bf21-b89723c3efb6-kube-api-access-58pn8\") pod \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.194511 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e43bef3b-68f9-4e72-bf21-b89723c3efb6-etc-swift\") pod \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.194588 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-dispersionconf\") pod \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.194622 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-ring-data-devices\") pod \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\" (UID: \"e43bef3b-68f9-4e72-bf21-b89723c3efb6\") " Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.195551 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e43bef3b-68f9-4e72-bf21-b89723c3efb6" (UID: "e43bef3b-68f9-4e72-bf21-b89723c3efb6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.195750 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e43bef3b-68f9-4e72-bf21-b89723c3efb6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e43bef3b-68f9-4e72-bf21-b89723c3efb6" (UID: "e43bef3b-68f9-4e72-bf21-b89723c3efb6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.201291 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43bef3b-68f9-4e72-bf21-b89723c3efb6-kube-api-access-58pn8" (OuterVolumeSpecName: "kube-api-access-58pn8") pod "e43bef3b-68f9-4e72-bf21-b89723c3efb6" (UID: "e43bef3b-68f9-4e72-bf21-b89723c3efb6"). InnerVolumeSpecName "kube-api-access-58pn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.222512 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-scripts" (OuterVolumeSpecName: "scripts") pod "e43bef3b-68f9-4e72-bf21-b89723c3efb6" (UID: "e43bef3b-68f9-4e72-bf21-b89723c3efb6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.224112 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e43bef3b-68f9-4e72-bf21-b89723c3efb6" (UID: "e43bef3b-68f9-4e72-bf21-b89723c3efb6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.224761 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e43bef3b-68f9-4e72-bf21-b89723c3efb6" (UID: "e43bef3b-68f9-4e72-bf21-b89723c3efb6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.297209 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.297253 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.297267 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e43bef3b-68f9-4e72-bf21-b89723c3efb6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.297279 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e43bef3b-68f9-4e72-bf21-b89723c3efb6-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.297292 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58pn8\" (UniqueName: \"kubernetes.io/projected/e43bef3b-68f9-4e72-bf21-b89723c3efb6-kube-api-access-58pn8\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.297306 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e43bef3b-68f9-4e72-bf21-b89723c3efb6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.698723 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cbb9745ca21d99ce4717cef6cdc68dac7b94fa5e4de579068fd54af7582e50b" Mar 09 10:01:17 crc kubenswrapper[4692]: I0309 10:01:17.698779 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-6blv4" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.083189 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43bef3b-68f9-4e72-bf21-b89723c3efb6" path="/var/lib/kubelet/pods/e43bef3b-68f9-4e72-bf21-b89723c3efb6/volumes" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.243417 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzh82"] Mar 09 10:01:18 crc kubenswrapper[4692]: E0309 10:01:18.243860 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43bef3b-68f9-4e72-bf21-b89723c3efb6" containerName="swift-ring-rebalance" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.243887 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43bef3b-68f9-4e72-bf21-b89723c3efb6" containerName="swift-ring-rebalance" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.244136 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="e43bef3b-68f9-4e72-bf21-b89723c3efb6" containerName="swift-ring-rebalance" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.244869 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.247093 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.250841 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.256413 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzh82"] Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.314400 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-scripts\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.314489 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-dispersionconf\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.314545 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c6cae175-2445-43a9-ae18-3f74cb266a4b-etc-swift\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.314644 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lnhq\" (UniqueName: \"kubernetes.io/projected/c6cae175-2445-43a9-ae18-3f74cb266a4b-kube-api-access-7lnhq\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.314797 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.314846 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-swiftconf\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.415740 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-dispersionconf\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.415809 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c6cae175-2445-43a9-ae18-3f74cb266a4b-etc-swift\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.415836 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lnhq\" (UniqueName: \"kubernetes.io/projected/c6cae175-2445-43a9-ae18-3f74cb266a4b-kube-api-access-7lnhq\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.415877 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.415896 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-swiftconf\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.415940 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-scripts\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.416685 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c6cae175-2445-43a9-ae18-3f74cb266a4b-etc-swift\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.416979 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-scripts\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.417132 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-ring-data-devices\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.420971 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-swiftconf\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.429302 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-dispersionconf\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.438377 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lnhq\" (UniqueName: \"kubernetes.io/projected/c6cae175-2445-43a9-ae18-3f74cb266a4b-kube-api-access-7lnhq\") pod \"swift-ring-rebalance-debug-gzh82\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.574950 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:18 crc kubenswrapper[4692]: I0309 10:01:18.834508 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzh82"] Mar 09 10:01:18 crc kubenswrapper[4692]: W0309 10:01:18.844450 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6cae175_2445_43a9_ae18_3f74cb266a4b.slice/crio-5563dce15441c7e5ebf12a411705c4cc65022cdf15d7c6ebc962d455a426c00e WatchSource:0}: Error finding container 5563dce15441c7e5ebf12a411705c4cc65022cdf15d7c6ebc962d455a426c00e: Status 404 returned error can't find the container with id 5563dce15441c7e5ebf12a411705c4cc65022cdf15d7c6ebc962d455a426c00e Mar 09 10:01:19 crc kubenswrapper[4692]: I0309 10:01:19.731476 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" event={"ID":"c6cae175-2445-43a9-ae18-3f74cb266a4b","Type":"ContainerStarted","Data":"770a4b8b6e13f687f65aa53102eee5c4e7ea702e73a0c2d74d586c71ca32055f"} Mar 09 10:01:19 crc kubenswrapper[4692]: I0309 10:01:19.731917 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" event={"ID":"c6cae175-2445-43a9-ae18-3f74cb266a4b","Type":"ContainerStarted","Data":"5563dce15441c7e5ebf12a411705c4cc65022cdf15d7c6ebc962d455a426c00e"} Mar 09 10:01:19 crc kubenswrapper[4692]: I0309 10:01:19.766518 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" podStartSLOduration=1.766494984 podStartE2EDuration="1.766494984s" podCreationTimestamp="2026-03-09 10:01:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:19.754636161 +0000 UTC m=+2480.579371782" watchObservedRunningTime="2026-03-09 10:01:19.766494984 +0000 UTC m=+2480.591230585" Mar 09 10:01:20 crc kubenswrapper[4692]: I0309 10:01:20.748891 4692 generic.go:334] "Generic (PLEG): container finished" podID="c6cae175-2445-43a9-ae18-3f74cb266a4b" containerID="770a4b8b6e13f687f65aa53102eee5c4e7ea702e73a0c2d74d586c71ca32055f" exitCode=0 Mar 09 10:01:20 crc kubenswrapper[4692]: I0309 10:01:20.748959 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" event={"ID":"c6cae175-2445-43a9-ae18-3f74cb266a4b","Type":"ContainerDied","Data":"770a4b8b6e13f687f65aa53102eee5c4e7ea702e73a0c2d74d586c71ca32055f"} Mar 09 10:01:21 crc kubenswrapper[4692]: I0309 10:01:21.071909 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:01:21 crc kubenswrapper[4692]: E0309 10:01:21.072241 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.051209 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.089357 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzh82"] Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.102553 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzh82"] Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.179090 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-scripts\") pod \"c6cae175-2445-43a9-ae18-3f74cb266a4b\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.179275 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-swiftconf\") pod \"c6cae175-2445-43a9-ae18-3f74cb266a4b\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.179297 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-dispersionconf\") pod \"c6cae175-2445-43a9-ae18-3f74cb266a4b\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.179353 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-ring-data-devices\") pod \"c6cae175-2445-43a9-ae18-3f74cb266a4b\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.179481 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c6cae175-2445-43a9-ae18-3f74cb266a4b-etc-swift\") pod \"c6cae175-2445-43a9-ae18-3f74cb266a4b\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.179529 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lnhq\" (UniqueName: \"kubernetes.io/projected/c6cae175-2445-43a9-ae18-3f74cb266a4b-kube-api-access-7lnhq\") pod \"c6cae175-2445-43a9-ae18-3f74cb266a4b\" (UID: \"c6cae175-2445-43a9-ae18-3f74cb266a4b\") " Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.180399 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c6cae175-2445-43a9-ae18-3f74cb266a4b" (UID: "c6cae175-2445-43a9-ae18-3f74cb266a4b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.180553 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6cae175-2445-43a9-ae18-3f74cb266a4b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c6cae175-2445-43a9-ae18-3f74cb266a4b" (UID: "c6cae175-2445-43a9-ae18-3f74cb266a4b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.186508 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6cae175-2445-43a9-ae18-3f74cb266a4b-kube-api-access-7lnhq" (OuterVolumeSpecName: "kube-api-access-7lnhq") pod "c6cae175-2445-43a9-ae18-3f74cb266a4b" (UID: "c6cae175-2445-43a9-ae18-3f74cb266a4b"). InnerVolumeSpecName "kube-api-access-7lnhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.205473 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c6cae175-2445-43a9-ae18-3f74cb266a4b" (UID: "c6cae175-2445-43a9-ae18-3f74cb266a4b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.205924 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c6cae175-2445-43a9-ae18-3f74cb266a4b" (UID: "c6cae175-2445-43a9-ae18-3f74cb266a4b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.214384 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-scripts" (OuterVolumeSpecName: "scripts") pod "c6cae175-2445-43a9-ae18-3f74cb266a4b" (UID: "c6cae175-2445-43a9-ae18-3f74cb266a4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.282228 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.282303 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c6cae175-2445-43a9-ae18-3f74cb266a4b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.282318 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lnhq\" (UniqueName: \"kubernetes.io/projected/c6cae175-2445-43a9-ae18-3f74cb266a4b-kube-api-access-7lnhq\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.282335 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c6cae175-2445-43a9-ae18-3f74cb266a4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.282350 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.282361 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c6cae175-2445-43a9-ae18-3f74cb266a4b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.773533 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5563dce15441c7e5ebf12a411705c4cc65022cdf15d7c6ebc962d455a426c00e" Mar 09 10:01:22 crc kubenswrapper[4692]: I0309 10:01:22.773617 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzh82" Mar 09 10:01:22 crc kubenswrapper[4692]: E0309 10:01:22.920869 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d86bcbf_f2c1_4253_b3bc_0b76d6bb23a2.slice/crio-da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f\": RecentStats: unable to find data in memory cache]" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.260634 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw"] Mar 09 10:01:23 crc kubenswrapper[4692]: E0309 10:01:23.261004 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6cae175-2445-43a9-ae18-3f74cb266a4b" containerName="swift-ring-rebalance" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.261019 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6cae175-2445-43a9-ae18-3f74cb266a4b" containerName="swift-ring-rebalance" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.261222 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6cae175-2445-43a9-ae18-3f74cb266a4b" containerName="swift-ring-rebalance" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.261770 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.264329 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.264690 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.275938 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw"] Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.401193 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-dispersionconf\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.401312 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqxwt\" (UniqueName: \"kubernetes.io/projected/850399eb-1ae0-442d-9e65-feecef92981b-kube-api-access-fqxwt\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.401387 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/850399eb-1ae0-442d-9e65-feecef92981b-etc-swift\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.401465 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-swiftconf\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.401505 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-ring-data-devices\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.401555 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-scripts\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.503220 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqxwt\" (UniqueName: \"kubernetes.io/projected/850399eb-1ae0-442d-9e65-feecef92981b-kube-api-access-fqxwt\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.503597 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/850399eb-1ae0-442d-9e65-feecef92981b-etc-swift\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.503709 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-swiftconf\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.503782 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-ring-data-devices\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.503884 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-scripts\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.503977 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-dispersionconf\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.505716 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/850399eb-1ae0-442d-9e65-feecef92981b-etc-swift\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.505841 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-ring-data-devices\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.506385 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-scripts\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.510976 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-swiftconf\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.511178 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-dispersionconf\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.530316 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqxwt\" (UniqueName: \"kubernetes.io/projected/850399eb-1ae0-442d-9e65-feecef92981b-kube-api-access-fqxwt\") pod \"swift-ring-rebalance-debug-vc7vw\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.586983 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:23 crc kubenswrapper[4692]: I0309 10:01:23.860408 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw"] Mar 09 10:01:24 crc kubenswrapper[4692]: I0309 10:01:24.082173 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6cae175-2445-43a9-ae18-3f74cb266a4b" path="/var/lib/kubelet/pods/c6cae175-2445-43a9-ae18-3f74cb266a4b/volumes" Mar 09 10:01:24 crc kubenswrapper[4692]: I0309 10:01:24.823019 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" event={"ID":"850399eb-1ae0-442d-9e65-feecef92981b","Type":"ContainerStarted","Data":"827267f37ac23a596a19c539b1f3aa69786724c2f77958aa18d7d2e6f3f232b8"} Mar 09 10:01:24 crc kubenswrapper[4692]: I0309 10:01:24.823549 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" event={"ID":"850399eb-1ae0-442d-9e65-feecef92981b","Type":"ContainerStarted","Data":"4a01cf56a96d802f67e395b2ad45e1f33f994ad018bea88ab98570954b261798"} Mar 09 10:01:24 crc kubenswrapper[4692]: I0309 10:01:24.853486 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" podStartSLOduration=1.853464805 podStartE2EDuration="1.853464805s" podCreationTimestamp="2026-03-09 10:01:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:24.851920021 +0000 UTC m=+2485.676655612" watchObservedRunningTime="2026-03-09 10:01:24.853464805 +0000 UTC m=+2485.678200386" Mar 09 10:01:25 crc kubenswrapper[4692]: I0309 10:01:25.836795 4692 generic.go:334] "Generic (PLEG): container finished" podID="850399eb-1ae0-442d-9e65-feecef92981b" containerID="827267f37ac23a596a19c539b1f3aa69786724c2f77958aa18d7d2e6f3f232b8" exitCode=0 Mar 09 10:01:25 crc kubenswrapper[4692]: I0309 10:01:25.836868 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" event={"ID":"850399eb-1ae0-442d-9e65-feecef92981b","Type":"ContainerDied","Data":"827267f37ac23a596a19c539b1f3aa69786724c2f77958aa18d7d2e6f3f232b8"} Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.143407 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.198119 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw"] Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.206568 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw"] Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.288038 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-scripts\") pod \"850399eb-1ae0-442d-9e65-feecef92981b\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.288481 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-swiftconf\") pod \"850399eb-1ae0-442d-9e65-feecef92981b\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.288711 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/850399eb-1ae0-442d-9e65-feecef92981b-etc-swift\") pod \"850399eb-1ae0-442d-9e65-feecef92981b\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.288818 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-ring-data-devices\") pod \"850399eb-1ae0-442d-9e65-feecef92981b\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.288967 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqxwt\" (UniqueName: \"kubernetes.io/projected/850399eb-1ae0-442d-9e65-feecef92981b-kube-api-access-fqxwt\") pod \"850399eb-1ae0-442d-9e65-feecef92981b\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.289102 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-dispersionconf\") pod \"850399eb-1ae0-442d-9e65-feecef92981b\" (UID: \"850399eb-1ae0-442d-9e65-feecef92981b\") " Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.289651 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "850399eb-1ae0-442d-9e65-feecef92981b" (UID: "850399eb-1ae0-442d-9e65-feecef92981b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.290765 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/850399eb-1ae0-442d-9e65-feecef92981b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "850399eb-1ae0-442d-9e65-feecef92981b" (UID: "850399eb-1ae0-442d-9e65-feecef92981b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.296797 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850399eb-1ae0-442d-9e65-feecef92981b-kube-api-access-fqxwt" (OuterVolumeSpecName: "kube-api-access-fqxwt") pod "850399eb-1ae0-442d-9e65-feecef92981b" (UID: "850399eb-1ae0-442d-9e65-feecef92981b"). InnerVolumeSpecName "kube-api-access-fqxwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.325403 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "850399eb-1ae0-442d-9e65-feecef92981b" (UID: "850399eb-1ae0-442d-9e65-feecef92981b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.328031 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-scripts" (OuterVolumeSpecName: "scripts") pod "850399eb-1ae0-442d-9e65-feecef92981b" (UID: "850399eb-1ae0-442d-9e65-feecef92981b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.331824 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "850399eb-1ae0-442d-9e65-feecef92981b" (UID: "850399eb-1ae0-442d-9e65-feecef92981b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.392152 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqxwt\" (UniqueName: \"kubernetes.io/projected/850399eb-1ae0-442d-9e65-feecef92981b-kube-api-access-fqxwt\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.392430 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.392543 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.392627 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/850399eb-1ae0-442d-9e65-feecef92981b-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.392847 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/850399eb-1ae0-442d-9e65-feecef92981b-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.392932 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/850399eb-1ae0-442d-9e65-feecef92981b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.859287 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a01cf56a96d802f67e395b2ad45e1f33f994ad018bea88ab98570954b261798" Mar 09 10:01:27 crc kubenswrapper[4692]: I0309 10:01:27.859371 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-vc7vw" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.082542 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850399eb-1ae0-442d-9e65-feecef92981b" path="/var/lib/kubelet/pods/850399eb-1ae0-442d-9e65-feecef92981b/volumes" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.347849 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f"] Mar 09 10:01:28 crc kubenswrapper[4692]: E0309 10:01:28.349737 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850399eb-1ae0-442d-9e65-feecef92981b" containerName="swift-ring-rebalance" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.349844 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="850399eb-1ae0-442d-9e65-feecef92981b" containerName="swift-ring-rebalance" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.350324 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="850399eb-1ae0-442d-9e65-feecef92981b" containerName="swift-ring-rebalance" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.351030 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.356777 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.357079 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.363137 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f"] Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.409336 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqvn2\" (UniqueName: \"kubernetes.io/projected/c365134f-edb2-4ca8-8429-455e694a7ff4-kube-api-access-nqvn2\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.409408 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-scripts\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.409437 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-dispersionconf\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.409467 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-ring-data-devices\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.409586 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c365134f-edb2-4ca8-8429-455e694a7ff4-etc-swift\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.409609 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-swiftconf\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.511338 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqvn2\" (UniqueName: \"kubernetes.io/projected/c365134f-edb2-4ca8-8429-455e694a7ff4-kube-api-access-nqvn2\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.511434 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-scripts\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.511483 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-dispersionconf\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.511533 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-ring-data-devices\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.511654 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c365134f-edb2-4ca8-8429-455e694a7ff4-etc-swift\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.511696 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-swiftconf\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.512588 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c365134f-edb2-4ca8-8429-455e694a7ff4-etc-swift\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.512692 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-ring-data-devices\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.512938 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-scripts\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.523627 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-dispersionconf\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.524718 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-swiftconf\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.536778 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqvn2\" (UniqueName: \"kubernetes.io/projected/c365134f-edb2-4ca8-8429-455e694a7ff4-kube-api-access-nqvn2\") pod \"swift-ring-rebalance-debug-f4n6f\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:28 crc kubenswrapper[4692]: I0309 10:01:28.674009 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:29 crc kubenswrapper[4692]: I0309 10:01:29.163667 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f"] Mar 09 10:01:29 crc kubenswrapper[4692]: I0309 10:01:29.884029 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" event={"ID":"c365134f-edb2-4ca8-8429-455e694a7ff4","Type":"ContainerStarted","Data":"84d8c48309a4d506c7678cd9f43585f04c95518a564ac198a40f544bdeef0325"} Mar 09 10:01:29 crc kubenswrapper[4692]: I0309 10:01:29.884493 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" event={"ID":"c365134f-edb2-4ca8-8429-455e694a7ff4","Type":"ContainerStarted","Data":"675fa4ef4fff0afc8c8116202ce497c5798c5edae2b8e663a793e8851a405638"} Mar 09 10:01:29 crc kubenswrapper[4692]: I0309 10:01:29.910913 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" podStartSLOduration=1.910888562 podStartE2EDuration="1.910888562s" podCreationTimestamp="2026-03-09 10:01:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:29.910365448 +0000 UTC m=+2490.735101049" watchObservedRunningTime="2026-03-09 10:01:29.910888562 +0000 UTC m=+2490.735624153" Mar 09 10:01:30 crc kubenswrapper[4692]: I0309 10:01:30.900856 4692 generic.go:334] "Generic (PLEG): container finished" podID="c365134f-edb2-4ca8-8429-455e694a7ff4" containerID="84d8c48309a4d506c7678cd9f43585f04c95518a564ac198a40f544bdeef0325" exitCode=0 Mar 09 10:01:30 crc kubenswrapper[4692]: I0309 10:01:30.900979 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" event={"ID":"c365134f-edb2-4ca8-8429-455e694a7ff4","Type":"ContainerDied","Data":"84d8c48309a4d506c7678cd9f43585f04c95518a564ac198a40f544bdeef0325"} Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.204542 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.255672 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f"] Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.263359 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f"] Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.282303 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqvn2\" (UniqueName: \"kubernetes.io/projected/c365134f-edb2-4ca8-8429-455e694a7ff4-kube-api-access-nqvn2\") pod \"c365134f-edb2-4ca8-8429-455e694a7ff4\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.282574 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-swiftconf\") pod \"c365134f-edb2-4ca8-8429-455e694a7ff4\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.282715 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-scripts\") pod \"c365134f-edb2-4ca8-8429-455e694a7ff4\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.283694 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-ring-data-devices\") pod \"c365134f-edb2-4ca8-8429-455e694a7ff4\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.283742 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c365134f-edb2-4ca8-8429-455e694a7ff4-etc-swift\") pod \"c365134f-edb2-4ca8-8429-455e694a7ff4\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.283826 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-dispersionconf\") pod \"c365134f-edb2-4ca8-8429-455e694a7ff4\" (UID: \"c365134f-edb2-4ca8-8429-455e694a7ff4\") " Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.285028 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c365134f-edb2-4ca8-8429-455e694a7ff4" (UID: "c365134f-edb2-4ca8-8429-455e694a7ff4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.286158 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c365134f-edb2-4ca8-8429-455e694a7ff4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c365134f-edb2-4ca8-8429-455e694a7ff4" (UID: "c365134f-edb2-4ca8-8429-455e694a7ff4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.290069 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c365134f-edb2-4ca8-8429-455e694a7ff4-kube-api-access-nqvn2" (OuterVolumeSpecName: "kube-api-access-nqvn2") pod "c365134f-edb2-4ca8-8429-455e694a7ff4" (UID: "c365134f-edb2-4ca8-8429-455e694a7ff4"). InnerVolumeSpecName "kube-api-access-nqvn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.307283 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-scripts" (OuterVolumeSpecName: "scripts") pod "c365134f-edb2-4ca8-8429-455e694a7ff4" (UID: "c365134f-edb2-4ca8-8429-455e694a7ff4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.309776 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c365134f-edb2-4ca8-8429-455e694a7ff4" (UID: "c365134f-edb2-4ca8-8429-455e694a7ff4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.319670 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c365134f-edb2-4ca8-8429-455e694a7ff4" (UID: "c365134f-edb2-4ca8-8429-455e694a7ff4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.386661 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqvn2\" (UniqueName: \"kubernetes.io/projected/c365134f-edb2-4ca8-8429-455e694a7ff4-kube-api-access-nqvn2\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.386713 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.386723 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.386734 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c365134f-edb2-4ca8-8429-455e694a7ff4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.386742 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c365134f-edb2-4ca8-8429-455e694a7ff4-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.386754 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c365134f-edb2-4ca8-8429-455e694a7ff4-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.924772 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="675fa4ef4fff0afc8c8116202ce497c5798c5edae2b8e663a793e8851a405638" Mar 09 10:01:32 crc kubenswrapper[4692]: I0309 10:01:32.924888 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-f4n6f" Mar 09 10:01:33 crc kubenswrapper[4692]: E0309 10:01:33.144929 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d86bcbf_f2c1_4253_b3bc_0b76d6bb23a2.slice/crio-da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f\": RecentStats: unable to find data in memory cache]" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.427527 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p8tld"] Mar 09 10:01:33 crc kubenswrapper[4692]: E0309 10:01:33.428497 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c365134f-edb2-4ca8-8429-455e694a7ff4" containerName="swift-ring-rebalance" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.428515 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c365134f-edb2-4ca8-8429-455e694a7ff4" containerName="swift-ring-rebalance" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.428670 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c365134f-edb2-4ca8-8429-455e694a7ff4" containerName="swift-ring-rebalance" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.429542 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.433938 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.433976 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.442248 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p8tld"] Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.613842 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15da6e3d-59ee-49c8-9ba8-c43327870dfa-etc-swift\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.613927 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-ring-data-devices\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.614664 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-scripts\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.614714 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96n7k\" (UniqueName: \"kubernetes.io/projected/15da6e3d-59ee-49c8-9ba8-c43327870dfa-kube-api-access-96n7k\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.614756 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-swiftconf\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.614814 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-dispersionconf\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.716292 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-scripts\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.716383 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96n7k\" (UniqueName: \"kubernetes.io/projected/15da6e3d-59ee-49c8-9ba8-c43327870dfa-kube-api-access-96n7k\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.716415 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-swiftconf\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.716439 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-dispersionconf\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.716475 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15da6e3d-59ee-49c8-9ba8-c43327870dfa-etc-swift\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.716519 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-ring-data-devices\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.717556 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15da6e3d-59ee-49c8-9ba8-c43327870dfa-etc-swift\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.717781 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-scripts\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.718267 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-ring-data-devices\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.721999 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-dispersionconf\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.722482 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-swiftconf\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.736299 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96n7k\" (UniqueName: \"kubernetes.io/projected/15da6e3d-59ee-49c8-9ba8-c43327870dfa-kube-api-access-96n7k\") pod \"swift-ring-rebalance-debug-p8tld\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:33 crc kubenswrapper[4692]: I0309 10:01:33.755861 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:34 crc kubenswrapper[4692]: I0309 10:01:34.077980 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:01:34 crc kubenswrapper[4692]: E0309 10:01:34.078687 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:01:34 crc kubenswrapper[4692]: I0309 10:01:34.088879 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c365134f-edb2-4ca8-8429-455e694a7ff4" path="/var/lib/kubelet/pods/c365134f-edb2-4ca8-8429-455e694a7ff4/volumes" Mar 09 10:01:34 crc kubenswrapper[4692]: I0309 10:01:34.198229 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p8tld"] Mar 09 10:01:34 crc kubenswrapper[4692]: I0309 10:01:34.959409 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" event={"ID":"15da6e3d-59ee-49c8-9ba8-c43327870dfa","Type":"ContainerStarted","Data":"6b5465a8edc1c16c3d97c7b025ede82fafc800af58001ecd555d91cf9c5786d6"} Mar 09 10:01:34 crc kubenswrapper[4692]: I0309 10:01:34.959818 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" event={"ID":"15da6e3d-59ee-49c8-9ba8-c43327870dfa","Type":"ContainerStarted","Data":"02ec970f8d4bae654b3c6bb77656c787b827997c2f5b0a491fc634e7d2c5f442"} Mar 09 10:01:34 crc kubenswrapper[4692]: I0309 10:01:34.987957 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" podStartSLOduration=1.9879357739999999 podStartE2EDuration="1.987935774s" podCreationTimestamp="2026-03-09 10:01:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:34.982519292 +0000 UTC m=+2495.807254883" watchObservedRunningTime="2026-03-09 10:01:34.987935774 +0000 UTC m=+2495.812671365" Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.883023 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hlht7"] Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.885858 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.897934 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hlht7"] Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.968602 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-utilities\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.968653 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-catalog-content\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.968756 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l56m7\" (UniqueName: \"kubernetes.io/projected/24a4d518-545c-43ba-8941-1a1ad95a96ba-kube-api-access-l56m7\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.971294 4692 generic.go:334] "Generic (PLEG): container finished" podID="15da6e3d-59ee-49c8-9ba8-c43327870dfa" containerID="6b5465a8edc1c16c3d97c7b025ede82fafc800af58001ecd555d91cf9c5786d6" exitCode=0 Mar 09 10:01:35 crc kubenswrapper[4692]: I0309 10:01:35.971372 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" event={"ID":"15da6e3d-59ee-49c8-9ba8-c43327870dfa","Type":"ContainerDied","Data":"6b5465a8edc1c16c3d97c7b025ede82fafc800af58001ecd555d91cf9c5786d6"} Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.070764 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l56m7\" (UniqueName: \"kubernetes.io/projected/24a4d518-545c-43ba-8941-1a1ad95a96ba-kube-api-access-l56m7\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.070876 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-utilities\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.070910 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-catalog-content\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.071474 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-catalog-content\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.072203 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-utilities\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.111836 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l56m7\" (UniqueName: \"kubernetes.io/projected/24a4d518-545c-43ba-8941-1a1ad95a96ba-kube-api-access-l56m7\") pod \"certified-operators-hlht7\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.214500 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.539446 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hlht7"] Mar 09 10:01:36 crc kubenswrapper[4692]: W0309 10:01:36.548880 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24a4d518_545c_43ba_8941_1a1ad95a96ba.slice/crio-f92053330418d69fab9fbcbddba3520e430d2cd9c7432f6cc3a2be930b43d7e4 WatchSource:0}: Error finding container f92053330418d69fab9fbcbddba3520e430d2cd9c7432f6cc3a2be930b43d7e4: Status 404 returned error can't find the container with id f92053330418d69fab9fbcbddba3520e430d2cd9c7432f6cc3a2be930b43d7e4 Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.984138 4692 generic.go:334] "Generic (PLEG): container finished" podID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerID="b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201" exitCode=0 Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.984211 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlht7" event={"ID":"24a4d518-545c-43ba-8941-1a1ad95a96ba","Type":"ContainerDied","Data":"b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201"} Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.984259 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlht7" event={"ID":"24a4d518-545c-43ba-8941-1a1ad95a96ba","Type":"ContainerStarted","Data":"f92053330418d69fab9fbcbddba3520e430d2cd9c7432f6cc3a2be930b43d7e4"} Mar 09 10:01:36 crc kubenswrapper[4692]: I0309 10:01:36.986575 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.310683 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.356117 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p8tld"] Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.362937 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-p8tld"] Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.505545 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15da6e3d-59ee-49c8-9ba8-c43327870dfa-etc-swift\") pod \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.505725 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-scripts\") pod \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.505767 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96n7k\" (UniqueName: \"kubernetes.io/projected/15da6e3d-59ee-49c8-9ba8-c43327870dfa-kube-api-access-96n7k\") pod \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.505793 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-swiftconf\") pod \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.505841 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-ring-data-devices\") pod \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.505901 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-dispersionconf\") pod \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\" (UID: \"15da6e3d-59ee-49c8-9ba8-c43327870dfa\") " Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.506759 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "15da6e3d-59ee-49c8-9ba8-c43327870dfa" (UID: "15da6e3d-59ee-49c8-9ba8-c43327870dfa"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.506799 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15da6e3d-59ee-49c8-9ba8-c43327870dfa-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "15da6e3d-59ee-49c8-9ba8-c43327870dfa" (UID: "15da6e3d-59ee-49c8-9ba8-c43327870dfa"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.514813 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15da6e3d-59ee-49c8-9ba8-c43327870dfa-kube-api-access-96n7k" (OuterVolumeSpecName: "kube-api-access-96n7k") pod "15da6e3d-59ee-49c8-9ba8-c43327870dfa" (UID: "15da6e3d-59ee-49c8-9ba8-c43327870dfa"). InnerVolumeSpecName "kube-api-access-96n7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.533799 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "15da6e3d-59ee-49c8-9ba8-c43327870dfa" (UID: "15da6e3d-59ee-49c8-9ba8-c43327870dfa"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.536007 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "15da6e3d-59ee-49c8-9ba8-c43327870dfa" (UID: "15da6e3d-59ee-49c8-9ba8-c43327870dfa"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.538734 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-scripts" (OuterVolumeSpecName: "scripts") pod "15da6e3d-59ee-49c8-9ba8-c43327870dfa" (UID: "15da6e3d-59ee-49c8-9ba8-c43327870dfa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.607280 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.607359 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.607371 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/15da6e3d-59ee-49c8-9ba8-c43327870dfa-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.607419 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/15da6e3d-59ee-49c8-9ba8-c43327870dfa-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.607433 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96n7k\" (UniqueName: \"kubernetes.io/projected/15da6e3d-59ee-49c8-9ba8-c43327870dfa-kube-api-access-96n7k\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:37 crc kubenswrapper[4692]: I0309 10:01:37.607444 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/15da6e3d-59ee-49c8-9ba8-c43327870dfa-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:38 crc kubenswrapper[4692]: I0309 10:01:38.001201 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlht7" event={"ID":"24a4d518-545c-43ba-8941-1a1ad95a96ba","Type":"ContainerStarted","Data":"ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00"} Mar 09 10:01:38 crc kubenswrapper[4692]: I0309 10:01:38.007327 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-p8tld" Mar 09 10:01:38 crc kubenswrapper[4692]: I0309 10:01:38.010374 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02ec970f8d4bae654b3c6bb77656c787b827997c2f5b0a491fc634e7d2c5f442" Mar 09 10:01:38 crc kubenswrapper[4692]: I0309 10:01:38.082244 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15da6e3d-59ee-49c8-9ba8-c43327870dfa" path="/var/lib/kubelet/pods/15da6e3d-59ee-49c8-9ba8-c43327870dfa/volumes" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.023660 4692 generic.go:334] "Generic (PLEG): container finished" podID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerID="ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00" exitCode=0 Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.023716 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlht7" event={"ID":"24a4d518-545c-43ba-8941-1a1ad95a96ba","Type":"ContainerDied","Data":"ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00"} Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.642122 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k"] Mar 09 10:01:39 crc kubenswrapper[4692]: E0309 10:01:39.642737 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15da6e3d-59ee-49c8-9ba8-c43327870dfa" containerName="swift-ring-rebalance" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.642766 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="15da6e3d-59ee-49c8-9ba8-c43327870dfa" containerName="swift-ring-rebalance" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.643004 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="15da6e3d-59ee-49c8-9ba8-c43327870dfa" containerName="swift-ring-rebalance" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.643636 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.646553 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.649441 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.656849 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-dispersionconf\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.656898 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-scripts\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.656924 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ef3e28-1b0a-4348-a2a5-e925270ed688-etc-swift\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.657027 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdqfs\" (UniqueName: \"kubernetes.io/projected/73ef3e28-1b0a-4348-a2a5-e925270ed688-kube-api-access-xdqfs\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.657057 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-ring-data-devices\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.657097 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-swiftconf\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.659119 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k"] Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.759669 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-swiftconf\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.759788 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-dispersionconf\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.759820 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-scripts\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.759842 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ef3e28-1b0a-4348-a2a5-e925270ed688-etc-swift\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.759986 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdqfs\" (UniqueName: \"kubernetes.io/projected/73ef3e28-1b0a-4348-a2a5-e925270ed688-kube-api-access-xdqfs\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.760042 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-ring-data-devices\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.760466 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ef3e28-1b0a-4348-a2a5-e925270ed688-etc-swift\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.761198 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-scripts\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.761329 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-ring-data-devices\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.766017 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-dispersionconf\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.768807 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-swiftconf\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.781753 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdqfs\" (UniqueName: \"kubernetes.io/projected/73ef3e28-1b0a-4348-a2a5-e925270ed688-kube-api-access-xdqfs\") pod \"swift-ring-rebalance-debug-lsj8k\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:39 crc kubenswrapper[4692]: I0309 10:01:39.977014 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:40 crc kubenswrapper[4692]: I0309 10:01:40.041259 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlht7" event={"ID":"24a4d518-545c-43ba-8941-1a1ad95a96ba","Type":"ContainerStarted","Data":"5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2"} Mar 09 10:01:40 crc kubenswrapper[4692]: I0309 10:01:40.073734 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hlht7" podStartSLOduration=2.373640839 podStartE2EDuration="5.073712171s" podCreationTimestamp="2026-03-09 10:01:35 +0000 UTC" firstStartedPulling="2026-03-09 10:01:36.986226834 +0000 UTC m=+2497.810962415" lastFinishedPulling="2026-03-09 10:01:39.686298166 +0000 UTC m=+2500.511033747" observedRunningTime="2026-03-09 10:01:40.064709317 +0000 UTC m=+2500.889444908" watchObservedRunningTime="2026-03-09 10:01:40.073712171 +0000 UTC m=+2500.898447752" Mar 09 10:01:40 crc kubenswrapper[4692]: I0309 10:01:40.465542 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k"] Mar 09 10:01:40 crc kubenswrapper[4692]: W0309 10:01:40.474565 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73ef3e28_1b0a_4348_a2a5_e925270ed688.slice/crio-fb318cc6ae84ded287c217c931562d51ab17033d9ee0f313aab5b376a6e17d68 WatchSource:0}: Error finding container fb318cc6ae84ded287c217c931562d51ab17033d9ee0f313aab5b376a6e17d68: Status 404 returned error can't find the container with id fb318cc6ae84ded287c217c931562d51ab17033d9ee0f313aab5b376a6e17d68 Mar 09 10:01:41 crc kubenswrapper[4692]: I0309 10:01:41.068858 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" event={"ID":"73ef3e28-1b0a-4348-a2a5-e925270ed688","Type":"ContainerStarted","Data":"b5d2badbb84fe7c91271064b90fb4c482072b23ac2f5f8eb8163510a17538b0d"} Mar 09 10:01:41 crc kubenswrapper[4692]: I0309 10:01:41.069364 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" event={"ID":"73ef3e28-1b0a-4348-a2a5-e925270ed688","Type":"ContainerStarted","Data":"fb318cc6ae84ded287c217c931562d51ab17033d9ee0f313aab5b376a6e17d68"} Mar 09 10:01:41 crc kubenswrapper[4692]: I0309 10:01:41.097392 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" podStartSLOduration=2.097363004 podStartE2EDuration="2.097363004s" podCreationTimestamp="2026-03-09 10:01:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:41.090965954 +0000 UTC m=+2501.915701545" watchObservedRunningTime="2026-03-09 10:01:41.097363004 +0000 UTC m=+2501.922098585" Mar 09 10:01:42 crc kubenswrapper[4692]: I0309 10:01:42.081710 4692 generic.go:334] "Generic (PLEG): container finished" podID="73ef3e28-1b0a-4348-a2a5-e925270ed688" containerID="b5d2badbb84fe7c91271064b90fb4c482072b23ac2f5f8eb8163510a17538b0d" exitCode=0 Mar 09 10:01:42 crc kubenswrapper[4692]: I0309 10:01:42.081761 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" event={"ID":"73ef3e28-1b0a-4348-a2a5-e925270ed688","Type":"ContainerDied","Data":"b5d2badbb84fe7c91271064b90fb4c482072b23ac2f5f8eb8163510a17538b0d"} Mar 09 10:01:43 crc kubenswrapper[4692]: E0309 10:01:43.403565 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d86bcbf_f2c1_4253_b3bc_0b76d6bb23a2.slice/crio-da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f\": RecentStats: unable to find data in memory cache]" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.403906 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.452965 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k"] Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.462073 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k"] Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.525908 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ef3e28-1b0a-4348-a2a5-e925270ed688-etc-swift\") pod \"73ef3e28-1b0a-4348-a2a5-e925270ed688\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.525970 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdqfs\" (UniqueName: \"kubernetes.io/projected/73ef3e28-1b0a-4348-a2a5-e925270ed688-kube-api-access-xdqfs\") pod \"73ef3e28-1b0a-4348-a2a5-e925270ed688\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.526022 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-scripts\") pod \"73ef3e28-1b0a-4348-a2a5-e925270ed688\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.526047 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-swiftconf\") pod \"73ef3e28-1b0a-4348-a2a5-e925270ed688\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.526063 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-dispersionconf\") pod \"73ef3e28-1b0a-4348-a2a5-e925270ed688\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.526180 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-ring-data-devices\") pod \"73ef3e28-1b0a-4348-a2a5-e925270ed688\" (UID: \"73ef3e28-1b0a-4348-a2a5-e925270ed688\") " Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.527320 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ef3e28-1b0a-4348-a2a5-e925270ed688-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "73ef3e28-1b0a-4348-a2a5-e925270ed688" (UID: "73ef3e28-1b0a-4348-a2a5-e925270ed688"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.527336 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "73ef3e28-1b0a-4348-a2a5-e925270ed688" (UID: "73ef3e28-1b0a-4348-a2a5-e925270ed688"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.535000 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ef3e28-1b0a-4348-a2a5-e925270ed688-kube-api-access-xdqfs" (OuterVolumeSpecName: "kube-api-access-xdqfs") pod "73ef3e28-1b0a-4348-a2a5-e925270ed688" (UID: "73ef3e28-1b0a-4348-a2a5-e925270ed688"). InnerVolumeSpecName "kube-api-access-xdqfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.551525 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-scripts" (OuterVolumeSpecName: "scripts") pod "73ef3e28-1b0a-4348-a2a5-e925270ed688" (UID: "73ef3e28-1b0a-4348-a2a5-e925270ed688"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.551981 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "73ef3e28-1b0a-4348-a2a5-e925270ed688" (UID: "73ef3e28-1b0a-4348-a2a5-e925270ed688"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.552434 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "73ef3e28-1b0a-4348-a2a5-e925270ed688" (UID: "73ef3e28-1b0a-4348-a2a5-e925270ed688"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.627374 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.627428 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/73ef3e28-1b0a-4348-a2a5-e925270ed688-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.627439 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdqfs\" (UniqueName: \"kubernetes.io/projected/73ef3e28-1b0a-4348-a2a5-e925270ed688-kube-api-access-xdqfs\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.627450 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73ef3e28-1b0a-4348-a2a5-e925270ed688-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.627459 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:43 crc kubenswrapper[4692]: I0309 10:01:43.627496 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/73ef3e28-1b0a-4348-a2a5-e925270ed688-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.083762 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ef3e28-1b0a-4348-a2a5-e925270ed688" path="/var/lib/kubelet/pods/73ef3e28-1b0a-4348-a2a5-e925270ed688/volumes" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.099772 4692 scope.go:117] "RemoveContainer" containerID="b5d2badbb84fe7c91271064b90fb4c482072b23ac2f5f8eb8163510a17538b0d" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.099859 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-lsj8k" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.596813 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w"] Mar 09 10:01:44 crc kubenswrapper[4692]: E0309 10:01:44.597276 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ef3e28-1b0a-4348-a2a5-e925270ed688" containerName="swift-ring-rebalance" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.597293 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ef3e28-1b0a-4348-a2a5-e925270ed688" containerName="swift-ring-rebalance" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.597465 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="73ef3e28-1b0a-4348-a2a5-e925270ed688" containerName="swift-ring-rebalance" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.598028 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.600757 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.601453 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.613628 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w"] Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.661686 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-scripts\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.661751 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-swiftconf\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.661779 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c3046528-19c3-4de8-8a46-bd38b0eeffc6-etc-swift\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.661799 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-dispersionconf\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.661841 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-ring-data-devices\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.661949 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wz82\" (UniqueName: \"kubernetes.io/projected/c3046528-19c3-4de8-8a46-bd38b0eeffc6-kube-api-access-7wz82\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.763830 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-scripts\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.764347 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-swiftconf\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.764465 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c3046528-19c3-4de8-8a46-bd38b0eeffc6-etc-swift\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.764575 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-ring-data-devices\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.764677 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-dispersionconf\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.764776 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wz82\" (UniqueName: \"kubernetes.io/projected/c3046528-19c3-4de8-8a46-bd38b0eeffc6-kube-api-access-7wz82\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.764937 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c3046528-19c3-4de8-8a46-bd38b0eeffc6-etc-swift\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.765313 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-ring-data-devices\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.765548 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-scripts\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.772755 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-dispersionconf\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.772943 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-swiftconf\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.786241 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wz82\" (UniqueName: \"kubernetes.io/projected/c3046528-19c3-4de8-8a46-bd38b0eeffc6-kube-api-access-7wz82\") pod \"swift-ring-rebalance-debug-g9w6w\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:44 crc kubenswrapper[4692]: I0309 10:01:44.922722 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:45 crc kubenswrapper[4692]: I0309 10:01:45.071598 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:01:45 crc kubenswrapper[4692]: E0309 10:01:45.072295 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:01:45 crc kubenswrapper[4692]: I0309 10:01:45.212420 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w"] Mar 09 10:01:46 crc kubenswrapper[4692]: I0309 10:01:46.126781 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" event={"ID":"c3046528-19c3-4de8-8a46-bd38b0eeffc6","Type":"ContainerStarted","Data":"a310a9938246877ac6786c3964684c3c472aaa879eb1e1d2e494f4cd396ed6b1"} Mar 09 10:01:46 crc kubenswrapper[4692]: I0309 10:01:46.127337 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" event={"ID":"c3046528-19c3-4de8-8a46-bd38b0eeffc6","Type":"ContainerStarted","Data":"04a6e93568dcc83e51dc0357e059585a0f3176efb1bbcf22ba3d9ff930dd8222"} Mar 09 10:01:46 crc kubenswrapper[4692]: I0309 10:01:46.153970 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" podStartSLOduration=2.153947299 podStartE2EDuration="2.153947299s" podCreationTimestamp="2026-03-09 10:01:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:46.150616696 +0000 UTC m=+2506.975352297" watchObservedRunningTime="2026-03-09 10:01:46.153947299 +0000 UTC m=+2506.978682880" Mar 09 10:01:46 crc kubenswrapper[4692]: I0309 10:01:46.215056 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:46 crc kubenswrapper[4692]: I0309 10:01:46.215125 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:46 crc kubenswrapper[4692]: I0309 10:01:46.274986 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:47 crc kubenswrapper[4692]: I0309 10:01:47.138054 4692 generic.go:334] "Generic (PLEG): container finished" podID="c3046528-19c3-4de8-8a46-bd38b0eeffc6" containerID="a310a9938246877ac6786c3964684c3c472aaa879eb1e1d2e494f4cd396ed6b1" exitCode=0 Mar 09 10:01:47 crc kubenswrapper[4692]: I0309 10:01:47.138136 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" event={"ID":"c3046528-19c3-4de8-8a46-bd38b0eeffc6","Type":"ContainerDied","Data":"a310a9938246877ac6786c3964684c3c472aaa879eb1e1d2e494f4cd396ed6b1"} Mar 09 10:01:47 crc kubenswrapper[4692]: I0309 10:01:47.186568 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:47 crc kubenswrapper[4692]: I0309 10:01:47.234689 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hlht7"] Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.422253 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.455119 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wz82\" (UniqueName: \"kubernetes.io/projected/c3046528-19c3-4de8-8a46-bd38b0eeffc6-kube-api-access-7wz82\") pod \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.455197 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-dispersionconf\") pod \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.455244 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-ring-data-devices\") pod \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.455273 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-swiftconf\") pod \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.455348 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-scripts\") pod \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.455387 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c3046528-19c3-4de8-8a46-bd38b0eeffc6-etc-swift\") pod \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\" (UID: \"c3046528-19c3-4de8-8a46-bd38b0eeffc6\") " Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.456492 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c3046528-19c3-4de8-8a46-bd38b0eeffc6" (UID: "c3046528-19c3-4de8-8a46-bd38b0eeffc6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.456554 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3046528-19c3-4de8-8a46-bd38b0eeffc6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c3046528-19c3-4de8-8a46-bd38b0eeffc6" (UID: "c3046528-19c3-4de8-8a46-bd38b0eeffc6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.463016 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3046528-19c3-4de8-8a46-bd38b0eeffc6-kube-api-access-7wz82" (OuterVolumeSpecName: "kube-api-access-7wz82") pod "c3046528-19c3-4de8-8a46-bd38b0eeffc6" (UID: "c3046528-19c3-4de8-8a46-bd38b0eeffc6"). InnerVolumeSpecName "kube-api-access-7wz82". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.471534 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w"] Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.477442 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w"] Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.480129 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-scripts" (OuterVolumeSpecName: "scripts") pod "c3046528-19c3-4de8-8a46-bd38b0eeffc6" (UID: "c3046528-19c3-4de8-8a46-bd38b0eeffc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.483942 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c3046528-19c3-4de8-8a46-bd38b0eeffc6" (UID: "c3046528-19c3-4de8-8a46-bd38b0eeffc6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.484675 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c3046528-19c3-4de8-8a46-bd38b0eeffc6" (UID: "c3046528-19c3-4de8-8a46-bd38b0eeffc6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.556606 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wz82\" (UniqueName: \"kubernetes.io/projected/c3046528-19c3-4de8-8a46-bd38b0eeffc6-kube-api-access-7wz82\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.556644 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.556655 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.556664 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c3046528-19c3-4de8-8a46-bd38b0eeffc6-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.556673 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3046528-19c3-4de8-8a46-bd38b0eeffc6-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:48 crc kubenswrapper[4692]: I0309 10:01:48.556681 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c3046528-19c3-4de8-8a46-bd38b0eeffc6-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.162126 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-g9w6w" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.162127 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04a6e93568dcc83e51dc0357e059585a0f3176efb1bbcf22ba3d9ff930dd8222" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.162301 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hlht7" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="registry-server" containerID="cri-o://5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2" gracePeriod=2 Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.567781 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.605749 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk"] Mar 09 10:01:49 crc kubenswrapper[4692]: E0309 10:01:49.606360 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3046528-19c3-4de8-8a46-bd38b0eeffc6" containerName="swift-ring-rebalance" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.606523 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3046528-19c3-4de8-8a46-bd38b0eeffc6" containerName="swift-ring-rebalance" Mar 09 10:01:49 crc kubenswrapper[4692]: E0309 10:01:49.606584 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="registry-server" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.606731 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="registry-server" Mar 09 10:01:49 crc kubenswrapper[4692]: E0309 10:01:49.606807 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="extract-content" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.606864 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="extract-content" Mar 09 10:01:49 crc kubenswrapper[4692]: E0309 10:01:49.606952 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="extract-utilities" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.607011 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="extract-utilities" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.607221 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerName="registry-server" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.607304 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3046528-19c3-4de8-8a46-bd38b0eeffc6" containerName="swift-ring-rebalance" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.607962 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.612149 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.612605 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.621744 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk"] Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.670569 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-utilities\") pod \"24a4d518-545c-43ba-8941-1a1ad95a96ba\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.670623 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-catalog-content\") pod \"24a4d518-545c-43ba-8941-1a1ad95a96ba\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.670656 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l56m7\" (UniqueName: \"kubernetes.io/projected/24a4d518-545c-43ba-8941-1a1ad95a96ba-kube-api-access-l56m7\") pod \"24a4d518-545c-43ba-8941-1a1ad95a96ba\" (UID: \"24a4d518-545c-43ba-8941-1a1ad95a96ba\") " Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.671627 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-utilities" (OuterVolumeSpecName: "utilities") pod "24a4d518-545c-43ba-8941-1a1ad95a96ba" (UID: "24a4d518-545c-43ba-8941-1a1ad95a96ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.676818 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24a4d518-545c-43ba-8941-1a1ad95a96ba-kube-api-access-l56m7" (OuterVolumeSpecName: "kube-api-access-l56m7") pod "24a4d518-545c-43ba-8941-1a1ad95a96ba" (UID: "24a4d518-545c-43ba-8941-1a1ad95a96ba"). InnerVolumeSpecName "kube-api-access-l56m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.729746 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24a4d518-545c-43ba-8941-1a1ad95a96ba" (UID: "24a4d518-545c-43ba-8941-1a1ad95a96ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.772687 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-ring-data-devices\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773013 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6c152c65-43e5-40c6-a13d-fc817e208e27-etc-swift\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773185 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-scripts\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773295 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-swiftconf\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773410 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-dispersionconf\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773522 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf6cv\" (UniqueName: \"kubernetes.io/projected/6c152c65-43e5-40c6-a13d-fc817e208e27-kube-api-access-lf6cv\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773824 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773906 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a4d518-545c-43ba-8941-1a1ad95a96ba-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.773982 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l56m7\" (UniqueName: \"kubernetes.io/projected/24a4d518-545c-43ba-8941-1a1ad95a96ba-kube-api-access-l56m7\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.875231 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-ring-data-devices\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.875747 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6c152c65-43e5-40c6-a13d-fc817e208e27-etc-swift\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.875775 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-scripts\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.875796 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-swiftconf\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.875828 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-dispersionconf\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.875852 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf6cv\" (UniqueName: \"kubernetes.io/projected/6c152c65-43e5-40c6-a13d-fc817e208e27-kube-api-access-lf6cv\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.876789 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-scripts\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.876789 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-ring-data-devices\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.876985 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6c152c65-43e5-40c6-a13d-fc817e208e27-etc-swift\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.880981 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-swiftconf\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.881212 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-dispersionconf\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.898153 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf6cv\" (UniqueName: \"kubernetes.io/projected/6c152c65-43e5-40c6-a13d-fc817e208e27-kube-api-access-lf6cv\") pod \"swift-ring-rebalance-debug-4q7zk\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:49 crc kubenswrapper[4692]: I0309 10:01:49.939232 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.083252 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3046528-19c3-4de8-8a46-bd38b0eeffc6" path="/var/lib/kubelet/pods/c3046528-19c3-4de8-8a46-bd38b0eeffc6/volumes" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.176662 4692 generic.go:334] "Generic (PLEG): container finished" podID="24a4d518-545c-43ba-8941-1a1ad95a96ba" containerID="5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2" exitCode=0 Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.176718 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlht7" event={"ID":"24a4d518-545c-43ba-8941-1a1ad95a96ba","Type":"ContainerDied","Data":"5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2"} Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.176760 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hlht7" event={"ID":"24a4d518-545c-43ba-8941-1a1ad95a96ba","Type":"ContainerDied","Data":"f92053330418d69fab9fbcbddba3520e430d2cd9c7432f6cc3a2be930b43d7e4"} Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.176788 4692 scope.go:117] "RemoveContainer" containerID="5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.176877 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hlht7" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.205751 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hlht7"] Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.207335 4692 scope.go:117] "RemoveContainer" containerID="ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.213153 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hlht7"] Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.227670 4692 scope.go:117] "RemoveContainer" containerID="b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.249472 4692 scope.go:117] "RemoveContainer" containerID="5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2" Mar 09 10:01:50 crc kubenswrapper[4692]: E0309 10:01:50.250062 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2\": container with ID starting with 5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2 not found: ID does not exist" containerID="5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.250127 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2"} err="failed to get container status \"5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2\": rpc error: code = NotFound desc = could not find container \"5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2\": container with ID starting with 5fa2f4020a8915b9d6dcad7682c8f9c5445c5e5bc9a2a89b4a87de0d164601a2 not found: ID does not exist" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.250179 4692 scope.go:117] "RemoveContainer" containerID="ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00" Mar 09 10:01:50 crc kubenswrapper[4692]: E0309 10:01:50.250708 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00\": container with ID starting with ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00 not found: ID does not exist" containerID="ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.250765 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00"} err="failed to get container status \"ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00\": rpc error: code = NotFound desc = could not find container \"ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00\": container with ID starting with ae67deb5ee0fbd050c092b5bf7840b3482d1a22469f9dbfba5aba9c4dd4f5a00 not found: ID does not exist" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.250815 4692 scope.go:117] "RemoveContainer" containerID="b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201" Mar 09 10:01:50 crc kubenswrapper[4692]: E0309 10:01:50.251295 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201\": container with ID starting with b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201 not found: ID does not exist" containerID="b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.251378 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201"} err="failed to get container status \"b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201\": rpc error: code = NotFound desc = could not find container \"b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201\": container with ID starting with b73d3b78784e069ee2ce25a738fc324fc5cf1c9bdac2be5295fc761a03b63201 not found: ID does not exist" Mar 09 10:01:50 crc kubenswrapper[4692]: I0309 10:01:50.352097 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk"] Mar 09 10:01:51 crc kubenswrapper[4692]: I0309 10:01:51.189359 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" event={"ID":"6c152c65-43e5-40c6-a13d-fc817e208e27","Type":"ContainerStarted","Data":"7da720bc2cb5c047aa50b837bdbda97fcab23df2f7d82492b3175e76c1a858e7"} Mar 09 10:01:51 crc kubenswrapper[4692]: I0309 10:01:51.189852 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" event={"ID":"6c152c65-43e5-40c6-a13d-fc817e208e27","Type":"ContainerStarted","Data":"081e94af4ca9dae3005143438fc9c91b91d0ace4ebb7c9e067e8543f914cb549"} Mar 09 10:01:51 crc kubenswrapper[4692]: I0309 10:01:51.218890 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" podStartSLOduration=2.218860739 podStartE2EDuration="2.218860739s" podCreationTimestamp="2026-03-09 10:01:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:51.214621709 +0000 UTC m=+2512.039357290" watchObservedRunningTime="2026-03-09 10:01:51.218860739 +0000 UTC m=+2512.043596330" Mar 09 10:01:52 crc kubenswrapper[4692]: I0309 10:01:52.084667 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24a4d518-545c-43ba-8941-1a1ad95a96ba" path="/var/lib/kubelet/pods/24a4d518-545c-43ba-8941-1a1ad95a96ba/volumes" Mar 09 10:01:52 crc kubenswrapper[4692]: I0309 10:01:52.203060 4692 generic.go:334] "Generic (PLEG): container finished" podID="6c152c65-43e5-40c6-a13d-fc817e208e27" containerID="7da720bc2cb5c047aa50b837bdbda97fcab23df2f7d82492b3175e76c1a858e7" exitCode=0 Mar 09 10:01:52 crc kubenswrapper[4692]: I0309 10:01:52.203106 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" event={"ID":"6c152c65-43e5-40c6-a13d-fc817e208e27","Type":"ContainerDied","Data":"7da720bc2cb5c047aa50b837bdbda97fcab23df2f7d82492b3175e76c1a858e7"} Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.526083 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.567034 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk"] Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.567927 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk"] Mar 09 10:01:53 crc kubenswrapper[4692]: E0309 10:01:53.627346 4692 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d86bcbf_f2c1_4253_b3bc_0b76d6bb23a2.slice/crio-da38dbaebf6759eb3e8486c554d4ae55ddc7f504404388c07372123f01e81b5f\": RecentStats: unable to find data in memory cache]" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.639149 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-dispersionconf\") pod \"6c152c65-43e5-40c6-a13d-fc817e208e27\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.639599 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf6cv\" (UniqueName: \"kubernetes.io/projected/6c152c65-43e5-40c6-a13d-fc817e208e27-kube-api-access-lf6cv\") pod \"6c152c65-43e5-40c6-a13d-fc817e208e27\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.639720 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-swiftconf\") pod \"6c152c65-43e5-40c6-a13d-fc817e208e27\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.639971 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-ring-data-devices\") pod \"6c152c65-43e5-40c6-a13d-fc817e208e27\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.640244 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6c152c65-43e5-40c6-a13d-fc817e208e27-etc-swift\") pod \"6c152c65-43e5-40c6-a13d-fc817e208e27\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.640417 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-scripts\") pod \"6c152c65-43e5-40c6-a13d-fc817e208e27\" (UID: \"6c152c65-43e5-40c6-a13d-fc817e208e27\") " Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.641240 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "6c152c65-43e5-40c6-a13d-fc817e208e27" (UID: "6c152c65-43e5-40c6-a13d-fc817e208e27"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.641528 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c152c65-43e5-40c6-a13d-fc817e208e27-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "6c152c65-43e5-40c6-a13d-fc817e208e27" (UID: "6c152c65-43e5-40c6-a13d-fc817e208e27"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.651263 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c152c65-43e5-40c6-a13d-fc817e208e27-kube-api-access-lf6cv" (OuterVolumeSpecName: "kube-api-access-lf6cv") pod "6c152c65-43e5-40c6-a13d-fc817e208e27" (UID: "6c152c65-43e5-40c6-a13d-fc817e208e27"). InnerVolumeSpecName "kube-api-access-lf6cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.662596 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-scripts" (OuterVolumeSpecName: "scripts") pod "6c152c65-43e5-40c6-a13d-fc817e208e27" (UID: "6c152c65-43e5-40c6-a13d-fc817e208e27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.665181 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "6c152c65-43e5-40c6-a13d-fc817e208e27" (UID: "6c152c65-43e5-40c6-a13d-fc817e208e27"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.665300 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "6c152c65-43e5-40c6-a13d-fc817e208e27" (UID: "6c152c65-43e5-40c6-a13d-fc817e208e27"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.742435 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.742475 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/6c152c65-43e5-40c6-a13d-fc817e208e27-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.742486 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c152c65-43e5-40c6-a13d-fc817e208e27-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.742495 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.742505 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf6cv\" (UniqueName: \"kubernetes.io/projected/6c152c65-43e5-40c6-a13d-fc817e208e27-kube-api-access-lf6cv\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:53 crc kubenswrapper[4692]: I0309 10:01:53.742515 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/6c152c65-43e5-40c6-a13d-fc817e208e27-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.085998 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c152c65-43e5-40c6-a13d-fc817e208e27" path="/var/lib/kubelet/pods/6c152c65-43e5-40c6-a13d-fc817e208e27/volumes" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.225554 4692 scope.go:117] "RemoveContainer" containerID="7da720bc2cb5c047aa50b837bdbda97fcab23df2f7d82492b3175e76c1a858e7" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.225695 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-4q7zk" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.722013 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l"] Mar 09 10:01:54 crc kubenswrapper[4692]: E0309 10:01:54.722498 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c152c65-43e5-40c6-a13d-fc817e208e27" containerName="swift-ring-rebalance" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.722518 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c152c65-43e5-40c6-a13d-fc817e208e27" containerName="swift-ring-rebalance" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.722705 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c152c65-43e5-40c6-a13d-fc817e208e27" containerName="swift-ring-rebalance" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.723425 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.727328 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.727568 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.745390 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l"] Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.865888 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-ring-data-devices\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.866330 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42jf9\" (UniqueName: \"kubernetes.io/projected/27db5a09-2def-4b6a-80c5-aebefebfaa2f-kube-api-access-42jf9\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.866428 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-swiftconf\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.866669 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-dispersionconf\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.866764 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27db5a09-2def-4b6a-80c5-aebefebfaa2f-etc-swift\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.866943 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-scripts\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.968857 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-ring-data-devices\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.968949 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42jf9\" (UniqueName: \"kubernetes.io/projected/27db5a09-2def-4b6a-80c5-aebefebfaa2f-kube-api-access-42jf9\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.968994 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-swiftconf\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.969069 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-dispersionconf\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.969109 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27db5a09-2def-4b6a-80c5-aebefebfaa2f-etc-swift\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.969190 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-scripts\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.969945 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27db5a09-2def-4b6a-80c5-aebefebfaa2f-etc-swift\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.970183 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-scripts\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.970197 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-ring-data-devices\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.980010 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-swiftconf\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.985430 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-dispersionconf\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:54 crc kubenswrapper[4692]: I0309 10:01:54.985929 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42jf9\" (UniqueName: \"kubernetes.io/projected/27db5a09-2def-4b6a-80c5-aebefebfaa2f-kube-api-access-42jf9\") pod \"swift-ring-rebalance-debug-kbs5l\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:55 crc kubenswrapper[4692]: I0309 10:01:55.053722 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:55 crc kubenswrapper[4692]: I0309 10:01:55.287867 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l"] Mar 09 10:01:56 crc kubenswrapper[4692]: I0309 10:01:56.260462 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" event={"ID":"27db5a09-2def-4b6a-80c5-aebefebfaa2f","Type":"ContainerStarted","Data":"3b246ebbc47316d3bec9ff2fd16ed0332af23549f429cd76dede64fcef70c878"} Mar 09 10:01:56 crc kubenswrapper[4692]: I0309 10:01:56.260800 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" event={"ID":"27db5a09-2def-4b6a-80c5-aebefebfaa2f","Type":"ContainerStarted","Data":"78c63fb96f2a9cc8763577b929b1ad72953f542f0aa8c8f93cb35baa134b6d34"} Mar 09 10:01:56 crc kubenswrapper[4692]: I0309 10:01:56.291367 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" podStartSLOduration=2.29133412 podStartE2EDuration="2.29133412s" podCreationTimestamp="2026-03-09 10:01:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:01:56.283061177 +0000 UTC m=+2517.107796768" watchObservedRunningTime="2026-03-09 10:01:56.29133412 +0000 UTC m=+2517.116069701" Mar 09 10:01:57 crc kubenswrapper[4692]: I0309 10:01:57.277230 4692 generic.go:334] "Generic (PLEG): container finished" podID="27db5a09-2def-4b6a-80c5-aebefebfaa2f" containerID="3b246ebbc47316d3bec9ff2fd16ed0332af23549f429cd76dede64fcef70c878" exitCode=0 Mar 09 10:01:57 crc kubenswrapper[4692]: I0309 10:01:57.277384 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" event={"ID":"27db5a09-2def-4b6a-80c5-aebefebfaa2f","Type":"ContainerDied","Data":"3b246ebbc47316d3bec9ff2fd16ed0332af23549f429cd76dede64fcef70c878"} Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.592078 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.627387 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l"] Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.634550 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l"] Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.643229 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-ring-data-devices\") pod \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.643387 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-dispersionconf\") pod \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.643476 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-swiftconf\") pod \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.643509 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27db5a09-2def-4b6a-80c5-aebefebfaa2f-etc-swift\") pod \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.643667 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42jf9\" (UniqueName: \"kubernetes.io/projected/27db5a09-2def-4b6a-80c5-aebefebfaa2f-kube-api-access-42jf9\") pod \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.643758 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-scripts\") pod \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\" (UID: \"27db5a09-2def-4b6a-80c5-aebefebfaa2f\") " Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.644494 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "27db5a09-2def-4b6a-80c5-aebefebfaa2f" (UID: "27db5a09-2def-4b6a-80c5-aebefebfaa2f"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.644723 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27db5a09-2def-4b6a-80c5-aebefebfaa2f-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "27db5a09-2def-4b6a-80c5-aebefebfaa2f" (UID: "27db5a09-2def-4b6a-80c5-aebefebfaa2f"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.645049 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.645066 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/27db5a09-2def-4b6a-80c5-aebefebfaa2f-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.671717 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27db5a09-2def-4b6a-80c5-aebefebfaa2f-kube-api-access-42jf9" (OuterVolumeSpecName: "kube-api-access-42jf9") pod "27db5a09-2def-4b6a-80c5-aebefebfaa2f" (UID: "27db5a09-2def-4b6a-80c5-aebefebfaa2f"). InnerVolumeSpecName "kube-api-access-42jf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.675314 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-scripts" (OuterVolumeSpecName: "scripts") pod "27db5a09-2def-4b6a-80c5-aebefebfaa2f" (UID: "27db5a09-2def-4b6a-80c5-aebefebfaa2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.690888 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "27db5a09-2def-4b6a-80c5-aebefebfaa2f" (UID: "27db5a09-2def-4b6a-80c5-aebefebfaa2f"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.698242 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "27db5a09-2def-4b6a-80c5-aebefebfaa2f" (UID: "27db5a09-2def-4b6a-80c5-aebefebfaa2f"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.747247 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42jf9\" (UniqueName: \"kubernetes.io/projected/27db5a09-2def-4b6a-80c5-aebefebfaa2f-kube-api-access-42jf9\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.747311 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/27db5a09-2def-4b6a-80c5-aebefebfaa2f-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.747326 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:58 crc kubenswrapper[4692]: I0309 10:01:58.747342 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/27db5a09-2def-4b6a-80c5-aebefebfaa2f-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.071774 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:01:59 crc kubenswrapper[4692]: E0309 10:01:59.072235 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.299418 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78c63fb96f2a9cc8763577b929b1ad72953f542f0aa8c8f93cb35baa134b6d34" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.299507 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-kbs5l" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.797258 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv"] Mar 09 10:01:59 crc kubenswrapper[4692]: E0309 10:01:59.797723 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27db5a09-2def-4b6a-80c5-aebefebfaa2f" containerName="swift-ring-rebalance" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.797743 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="27db5a09-2def-4b6a-80c5-aebefebfaa2f" containerName="swift-ring-rebalance" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.797941 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="27db5a09-2def-4b6a-80c5-aebefebfaa2f" containerName="swift-ring-rebalance" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.798597 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv"] Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.798700 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.823635 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.823826 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.866569 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-swiftconf\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.867035 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-dispersionconf\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.867073 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e15d560-cd47-48f3-98ef-bda5d74ba366-etc-swift\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.867111 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-ring-data-devices\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.867140 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-scripts\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.867354 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77x2t\" (UniqueName: \"kubernetes.io/projected/5e15d560-cd47-48f3-98ef-bda5d74ba366-kube-api-access-77x2t\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.969334 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-ring-data-devices\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.969398 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-scripts\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.969433 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77x2t\" (UniqueName: \"kubernetes.io/projected/5e15d560-cd47-48f3-98ef-bda5d74ba366-kube-api-access-77x2t\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.969519 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-swiftconf\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.969547 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-dispersionconf\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.969575 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e15d560-cd47-48f3-98ef-bda5d74ba366-etc-swift\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.970205 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e15d560-cd47-48f3-98ef-bda5d74ba366-etc-swift\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.970451 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-ring-data-devices\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.970714 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-scripts\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.976768 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-swiftconf\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.977124 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-dispersionconf\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:01:59 crc kubenswrapper[4692]: I0309 10:01:59.988870 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77x2t\" (UniqueName: \"kubernetes.io/projected/5e15d560-cd47-48f3-98ef-bda5d74ba366-kube-api-access-77x2t\") pod \"swift-ring-rebalance-debug-gzbsv\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.085231 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27db5a09-2def-4b6a-80c5-aebefebfaa2f" path="/var/lib/kubelet/pods/27db5a09-2def-4b6a-80c5-aebefebfaa2f/volumes" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.141226 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550842-jtclt"] Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.142331 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550842-jtclt" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.145095 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.145170 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.145389 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.146037 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.159665 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550842-jtclt"] Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.171705 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5kgd\" (UniqueName: \"kubernetes.io/projected/536ee033-b615-4af9-b26e-81da38bdc053-kube-api-access-g5kgd\") pod \"auto-csr-approver-29550842-jtclt\" (UID: \"536ee033-b615-4af9-b26e-81da38bdc053\") " pod="openshift-infra/auto-csr-approver-29550842-jtclt" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.275769 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5kgd\" (UniqueName: \"kubernetes.io/projected/536ee033-b615-4af9-b26e-81da38bdc053-kube-api-access-g5kgd\") pod \"auto-csr-approver-29550842-jtclt\" (UID: \"536ee033-b615-4af9-b26e-81da38bdc053\") " pod="openshift-infra/auto-csr-approver-29550842-jtclt" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.308591 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5kgd\" (UniqueName: \"kubernetes.io/projected/536ee033-b615-4af9-b26e-81da38bdc053-kube-api-access-g5kgd\") pod \"auto-csr-approver-29550842-jtclt\" (UID: \"536ee033-b615-4af9-b26e-81da38bdc053\") " pod="openshift-infra/auto-csr-approver-29550842-jtclt" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.470974 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550842-jtclt" Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.600129 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv"] Mar 09 10:02:00 crc kubenswrapper[4692]: I0309 10:02:00.912704 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550842-jtclt"] Mar 09 10:02:00 crc kubenswrapper[4692]: W0309 10:02:00.916861 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod536ee033_b615_4af9_b26e_81da38bdc053.slice/crio-db34d5b02564c4f3e8735b0365423a37cce9e9e1424325bb20cd3a0a55b2ad5c WatchSource:0}: Error finding container db34d5b02564c4f3e8735b0365423a37cce9e9e1424325bb20cd3a0a55b2ad5c: Status 404 returned error can't find the container with id db34d5b02564c4f3e8735b0365423a37cce9e9e1424325bb20cd3a0a55b2ad5c Mar 09 10:02:01 crc kubenswrapper[4692]: I0309 10:02:01.332685 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550842-jtclt" event={"ID":"536ee033-b615-4af9-b26e-81da38bdc053","Type":"ContainerStarted","Data":"db34d5b02564c4f3e8735b0365423a37cce9e9e1424325bb20cd3a0a55b2ad5c"} Mar 09 10:02:01 crc kubenswrapper[4692]: I0309 10:02:01.334213 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" event={"ID":"5e15d560-cd47-48f3-98ef-bda5d74ba366","Type":"ContainerStarted","Data":"841a06afb4f2229936e234a6d05cb5713cbb0314030192ff58e71398fc14bd14"} Mar 09 10:02:01 crc kubenswrapper[4692]: I0309 10:02:01.334238 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" event={"ID":"5e15d560-cd47-48f3-98ef-bda5d74ba366","Type":"ContainerStarted","Data":"fc9340f91dd03898ef1e925ab4ca850e62559c60cbfcc2f9f3b28bb24393239a"} Mar 09 10:02:01 crc kubenswrapper[4692]: I0309 10:02:01.358562 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" podStartSLOduration=2.358529033 podStartE2EDuration="2.358529033s" podCreationTimestamp="2026-03-09 10:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-09 10:02:01.352334179 +0000 UTC m=+2522.177069780" watchObservedRunningTime="2026-03-09 10:02:01.358529033 +0000 UTC m=+2522.183264614" Mar 09 10:02:02 crc kubenswrapper[4692]: I0309 10:02:02.355059 4692 generic.go:334] "Generic (PLEG): container finished" podID="5e15d560-cd47-48f3-98ef-bda5d74ba366" containerID="841a06afb4f2229936e234a6d05cb5713cbb0314030192ff58e71398fc14bd14" exitCode=0 Mar 09 10:02:02 crc kubenswrapper[4692]: I0309 10:02:02.355181 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" event={"ID":"5e15d560-cd47-48f3-98ef-bda5d74ba366","Type":"ContainerDied","Data":"841a06afb4f2229936e234a6d05cb5713cbb0314030192ff58e71398fc14bd14"} Mar 09 10:02:02 crc kubenswrapper[4692]: I0309 10:02:02.358585 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550842-jtclt" event={"ID":"536ee033-b615-4af9-b26e-81da38bdc053","Type":"ContainerStarted","Data":"effa3c186bf8f4ecb1a079539c061fca884e09581bc30f871f87090bacf83ea2"} Mar 09 10:02:02 crc kubenswrapper[4692]: I0309 10:02:02.396116 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29550842-jtclt" podStartSLOduration=1.2862946800000001 podStartE2EDuration="2.396074678s" podCreationTimestamp="2026-03-09 10:02:00 +0000 UTC" firstStartedPulling="2026-03-09 10:02:00.919583538 +0000 UTC m=+2521.744319119" lastFinishedPulling="2026-03-09 10:02:02.029363546 +0000 UTC m=+2522.854099117" observedRunningTime="2026-03-09 10:02:02.393071153 +0000 UTC m=+2523.217806734" watchObservedRunningTime="2026-03-09 10:02:02.396074678 +0000 UTC m=+2523.220810269" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.370901 4692 generic.go:334] "Generic (PLEG): container finished" podID="536ee033-b615-4af9-b26e-81da38bdc053" containerID="effa3c186bf8f4ecb1a079539c061fca884e09581bc30f871f87090bacf83ea2" exitCode=0 Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.370966 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550842-jtclt" event={"ID":"536ee033-b615-4af9-b26e-81da38bdc053","Type":"ContainerDied","Data":"effa3c186bf8f4ecb1a079539c061fca884e09581bc30f871f87090bacf83ea2"} Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.656531 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.699683 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv"] Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.707880 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv"] Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.831361 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77x2t\" (UniqueName: \"kubernetes.io/projected/5e15d560-cd47-48f3-98ef-bda5d74ba366-kube-api-access-77x2t\") pod \"5e15d560-cd47-48f3-98ef-bda5d74ba366\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.831549 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-scripts\") pod \"5e15d560-cd47-48f3-98ef-bda5d74ba366\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.831590 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-swiftconf\") pod \"5e15d560-cd47-48f3-98ef-bda5d74ba366\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.831625 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e15d560-cd47-48f3-98ef-bda5d74ba366-etc-swift\") pod \"5e15d560-cd47-48f3-98ef-bda5d74ba366\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.831686 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-dispersionconf\") pod \"5e15d560-cd47-48f3-98ef-bda5d74ba366\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.831709 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-ring-data-devices\") pod \"5e15d560-cd47-48f3-98ef-bda5d74ba366\" (UID: \"5e15d560-cd47-48f3-98ef-bda5d74ba366\") " Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.833439 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5e15d560-cd47-48f3-98ef-bda5d74ba366" (UID: "5e15d560-cd47-48f3-98ef-bda5d74ba366"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.833442 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e15d560-cd47-48f3-98ef-bda5d74ba366-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5e15d560-cd47-48f3-98ef-bda5d74ba366" (UID: "5e15d560-cd47-48f3-98ef-bda5d74ba366"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.842584 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e15d560-cd47-48f3-98ef-bda5d74ba366-kube-api-access-77x2t" (OuterVolumeSpecName: "kube-api-access-77x2t") pod "5e15d560-cd47-48f3-98ef-bda5d74ba366" (UID: "5e15d560-cd47-48f3-98ef-bda5d74ba366"). InnerVolumeSpecName "kube-api-access-77x2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.863565 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-scripts" (OuterVolumeSpecName: "scripts") pod "5e15d560-cd47-48f3-98ef-bda5d74ba366" (UID: "5e15d560-cd47-48f3-98ef-bda5d74ba366"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.863868 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5e15d560-cd47-48f3-98ef-bda5d74ba366" (UID: "5e15d560-cd47-48f3-98ef-bda5d74ba366"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.867329 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5e15d560-cd47-48f3-98ef-bda5d74ba366" (UID: "5e15d560-cd47-48f3-98ef-bda5d74ba366"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.934087 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77x2t\" (UniqueName: \"kubernetes.io/projected/5e15d560-cd47-48f3-98ef-bda5d74ba366-kube-api-access-77x2t\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.934136 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.934146 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.934180 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e15d560-cd47-48f3-98ef-bda5d74ba366-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.934191 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e15d560-cd47-48f3-98ef-bda5d74ba366-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:03 crc kubenswrapper[4692]: I0309 10:02:03.934202 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e15d560-cd47-48f3-98ef-bda5d74ba366-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.082596 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e15d560-cd47-48f3-98ef-bda5d74ba366" path="/var/lib/kubelet/pods/5e15d560-cd47-48f3-98ef-bda5d74ba366/volumes" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.383805 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-gzbsv" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.383798 4692 scope.go:117] "RemoveContainer" containerID="841a06afb4f2229936e234a6d05cb5713cbb0314030192ff58e71398fc14bd14" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.651186 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550842-jtclt" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.846263 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5kgd\" (UniqueName: \"kubernetes.io/projected/536ee033-b615-4af9-b26e-81da38bdc053-kube-api-access-g5kgd\") pod \"536ee033-b615-4af9-b26e-81da38bdc053\" (UID: \"536ee033-b615-4af9-b26e-81da38bdc053\") " Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.851382 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536ee033-b615-4af9-b26e-81da38bdc053-kube-api-access-g5kgd" (OuterVolumeSpecName: "kube-api-access-g5kgd") pod "536ee033-b615-4af9-b26e-81da38bdc053" (UID: "536ee033-b615-4af9-b26e-81da38bdc053"). InnerVolumeSpecName "kube-api-access-g5kgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.880041 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbn76"] Mar 09 10:02:04 crc kubenswrapper[4692]: E0309 10:02:04.880418 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536ee033-b615-4af9-b26e-81da38bdc053" containerName="oc" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.880432 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="536ee033-b615-4af9-b26e-81da38bdc053" containerName="oc" Mar 09 10:02:04 crc kubenswrapper[4692]: E0309 10:02:04.880450 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e15d560-cd47-48f3-98ef-bda5d74ba366" containerName="swift-ring-rebalance" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.880458 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e15d560-cd47-48f3-98ef-bda5d74ba366" containerName="swift-ring-rebalance" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.880614 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e15d560-cd47-48f3-98ef-bda5d74ba366" containerName="swift-ring-rebalance" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.880640 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="536ee033-b615-4af9-b26e-81da38bdc053" containerName="oc" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.881123 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.884642 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-scripts" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.888371 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"swift-kuttl-tests"/"swift-ring-config-data" Mar 09 10:02:04 crc kubenswrapper[4692]: I0309 10:02:04.896773 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbn76"] Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.362078 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5kgd\" (UniqueName: \"kubernetes.io/projected/536ee033-b615-4af9-b26e-81da38bdc053-kube-api-access-g5kgd\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.425623 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550842-jtclt" event={"ID":"536ee033-b615-4af9-b26e-81da38bdc053","Type":"ContainerDied","Data":"db34d5b02564c4f3e8735b0365423a37cce9e9e1424325bb20cd3a0a55b2ad5c"} Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.425671 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db34d5b02564c4f3e8735b0365423a37cce9e9e1424325bb20cd3a0a55b2ad5c" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.425708 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550842-jtclt" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.463281 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-swiftconf\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.463367 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-dispersionconf\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.463447 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-scripts\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.463567 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-ring-data-devices\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.463623 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dks59\" (UniqueName: \"kubernetes.io/projected/5e2813f5-c6e8-4db0-bc8c-84d69500306a-kube-api-access-dks59\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.463702 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e2813f5-c6e8-4db0-bc8c-84d69500306a-etc-swift\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.564709 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-ring-data-devices\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.564784 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dks59\" (UniqueName: \"kubernetes.io/projected/5e2813f5-c6e8-4db0-bc8c-84d69500306a-kube-api-access-dks59\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.564842 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e2813f5-c6e8-4db0-bc8c-84d69500306a-etc-swift\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.564896 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-swiftconf\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.564951 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-dispersionconf\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.564976 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-scripts\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.565669 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e2813f5-c6e8-4db0-bc8c-84d69500306a-etc-swift\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.565899 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-ring-data-devices\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.566377 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-scripts\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.581358 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-swiftconf\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.581617 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-dispersionconf\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.584325 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dks59\" (UniqueName: \"kubernetes.io/projected/5e2813f5-c6e8-4db0-bc8c-84d69500306a-kube-api-access-dks59\") pod \"swift-ring-rebalance-debug-pbn76\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.716030 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550836-glwp9"] Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.722930 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550836-glwp9"] Mar 09 10:02:05 crc kubenswrapper[4692]: I0309 10:02:05.823253 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.082731 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6d3a680-5cd9-487b-a928-b83c4f24728c" path="/var/lib/kubelet/pods/a6d3a680-5cd9-487b-a928-b83c4f24728c/volumes" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.260259 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbn76"] Mar 09 10:02:06 crc kubenswrapper[4692]: W0309 10:02:06.270081 4692 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e2813f5_c6e8_4db0_bc8c_84d69500306a.slice/crio-2afb912c69bfc24dfa056b3e0bf1b6d0a83e96b1c496115d1f5ede6719693e4a WatchSource:0}: Error finding container 2afb912c69bfc24dfa056b3e0bf1b6d0a83e96b1c496115d1f5ede6719693e4a: Status 404 returned error can't find the container with id 2afb912c69bfc24dfa056b3e0bf1b6d0a83e96b1c496115d1f5ede6719693e4a Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.436026 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" event={"ID":"5e2813f5-c6e8-4db0-bc8c-84d69500306a","Type":"ContainerStarted","Data":"2afb912c69bfc24dfa056b3e0bf1b6d0a83e96b1c496115d1f5ede6719693e4a"} Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.692310 4692 scope.go:117] "RemoveContainer" containerID="ba9798dc7eb1d2d2762cdb3f56cc3dabb328567811fe9aa71d948ce046363af8" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.721354 4692 scope.go:117] "RemoveContainer" containerID="a0d05f3c2440ecd9349096613efdff0e763516fc23ce41736cde0d954964667c" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.757404 4692 scope.go:117] "RemoveContainer" containerID="c9f7949195aeeb4133ea50e703e4d65a96217257e14b663b9e7f27e475897099" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.788017 4692 scope.go:117] "RemoveContainer" containerID="9f5573fdeb9526abf86dd80c7229949049acea80c0276a84dda1af12eb477ec8" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.840995 4692 scope.go:117] "RemoveContainer" containerID="145a219aa7ba64540dc0eb3092045f87edb0172d48d2a2b4ad39834c066a3381" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.873103 4692 scope.go:117] "RemoveContainer" containerID="31882b07d7ba81dbdf1376ed84debb308679d145ebcc592ffb10507fd63ac4a4" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.905062 4692 scope.go:117] "RemoveContainer" containerID="465939265c687ecb225c88675a6dc7746b4fa99f66577d802a7266bed10040ef" Mar 09 10:02:06 crc kubenswrapper[4692]: I0309 10:02:06.939303 4692 scope.go:117] "RemoveContainer" containerID="d1858c5b055955c720712dbcb9b715bd829de6079b9294b70e909c284eca4085" Mar 09 10:02:07 crc kubenswrapper[4692]: I0309 10:02:07.447012 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" event={"ID":"5e2813f5-c6e8-4db0-bc8c-84d69500306a","Type":"ContainerStarted","Data":"d3a75bc4ec744ff439c29abb826f12a24bd67721b8958526bfbaed07d073fbef"} Mar 09 10:02:08 crc kubenswrapper[4692]: I0309 10:02:08.459854 4692 generic.go:334] "Generic (PLEG): container finished" podID="5e2813f5-c6e8-4db0-bc8c-84d69500306a" containerID="d3a75bc4ec744ff439c29abb826f12a24bd67721b8958526bfbaed07d073fbef" exitCode=0 Mar 09 10:02:08 crc kubenswrapper[4692]: I0309 10:02:08.459906 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" event={"ID":"5e2813f5-c6e8-4db0-bc8c-84d69500306a","Type":"ContainerDied","Data":"d3a75bc4ec744ff439c29abb826f12a24bd67721b8958526bfbaed07d073fbef"} Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.771338 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.835054 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbn76"] Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.836659 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e2813f5-c6e8-4db0-bc8c-84d69500306a-etc-swift\") pod \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.836725 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-ring-data-devices\") pod \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.836789 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dks59\" (UniqueName: \"kubernetes.io/projected/5e2813f5-c6e8-4db0-bc8c-84d69500306a-kube-api-access-dks59\") pod \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.836852 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-scripts\") pod \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.836900 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-swiftconf\") pod \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.837011 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-dispersionconf\") pod \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\" (UID: \"5e2813f5-c6e8-4db0-bc8c-84d69500306a\") " Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.837974 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2813f5-c6e8-4db0-bc8c-84d69500306a-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "5e2813f5-c6e8-4db0-bc8c-84d69500306a" (UID: "5e2813f5-c6e8-4db0-bc8c-84d69500306a"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.838014 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "5e2813f5-c6e8-4db0-bc8c-84d69500306a" (UID: "5e2813f5-c6e8-4db0-bc8c-84d69500306a"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.842948 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["swift-kuttl-tests/swift-ring-rebalance-debug-pbn76"] Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.861127 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-scripts" (OuterVolumeSpecName: "scripts") pod "5e2813f5-c6e8-4db0-bc8c-84d69500306a" (UID: "5e2813f5-c6e8-4db0-bc8c-84d69500306a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.857628 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2813f5-c6e8-4db0-bc8c-84d69500306a-kube-api-access-dks59" (OuterVolumeSpecName: "kube-api-access-dks59") pod "5e2813f5-c6e8-4db0-bc8c-84d69500306a" (UID: "5e2813f5-c6e8-4db0-bc8c-84d69500306a"). InnerVolumeSpecName "kube-api-access-dks59". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.868014 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "5e2813f5-c6e8-4db0-bc8c-84d69500306a" (UID: "5e2813f5-c6e8-4db0-bc8c-84d69500306a"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.872524 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "5e2813f5-c6e8-4db0-bc8c-84d69500306a" (UID: "5e2813f5-c6e8-4db0-bc8c-84d69500306a"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.938712 4692 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-scripts\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.938751 4692 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.938764 4692 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/5e2813f5-c6e8-4db0-bc8c-84d69500306a-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.938777 4692 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/5e2813f5-c6e8-4db0-bc8c-84d69500306a-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.938790 4692 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/5e2813f5-c6e8-4db0-bc8c-84d69500306a-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:09 crc kubenswrapper[4692]: I0309 10:02:09.938799 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dks59\" (UniqueName: \"kubernetes.io/projected/5e2813f5-c6e8-4db0-bc8c-84d69500306a-kube-api-access-dks59\") on node \"crc\" DevicePath \"\"" Mar 09 10:02:10 crc kubenswrapper[4692]: I0309 10:02:10.081410 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2813f5-c6e8-4db0-bc8c-84d69500306a" path="/var/lib/kubelet/pods/5e2813f5-c6e8-4db0-bc8c-84d69500306a/volumes" Mar 09 10:02:10 crc kubenswrapper[4692]: I0309 10:02:10.482926 4692 scope.go:117] "RemoveContainer" containerID="d3a75bc4ec744ff439c29abb826f12a24bd67721b8958526bfbaed07d073fbef" Mar 09 10:02:10 crc kubenswrapper[4692]: I0309 10:02:10.482983 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="swift-kuttl-tests/swift-ring-rebalance-debug-pbn76" Mar 09 10:02:13 crc kubenswrapper[4692]: I0309 10:02:13.072258 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:02:13 crc kubenswrapper[4692]: E0309 10:02:13.072828 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:02:25 crc kubenswrapper[4692]: I0309 10:02:25.072625 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:02:25 crc kubenswrapper[4692]: I0309 10:02:25.640036 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"0c4aca92fab9686fc64015cb64a831b1e43d53698f0ab8c3789ce9a9e0a536fa"} Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.665537 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-t2242/must-gather-xtjbn"] Mar 09 10:02:32 crc kubenswrapper[4692]: E0309 10:02:32.666546 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2813f5-c6e8-4db0-bc8c-84d69500306a" containerName="swift-ring-rebalance" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.666561 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2813f5-c6e8-4db0-bc8c-84d69500306a" containerName="swift-ring-rebalance" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.666750 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2813f5-c6e8-4db0-bc8c-84d69500306a" containerName="swift-ring-rebalance" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.667666 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.673799 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-t2242"/"default-dockercfg-zbd56" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.675053 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t2242"/"openshift-service-ca.crt" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.676682 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-t2242"/"kube-root-ca.crt" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.682502 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t2242/must-gather-xtjbn"] Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.721566 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b0f0035-f0b8-48e7-86e8-68187ac429f0-must-gather-output\") pod \"must-gather-xtjbn\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.721703 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwsld\" (UniqueName: \"kubernetes.io/projected/6b0f0035-f0b8-48e7-86e8-68187ac429f0-kube-api-access-dwsld\") pod \"must-gather-xtjbn\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.823324 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwsld\" (UniqueName: \"kubernetes.io/projected/6b0f0035-f0b8-48e7-86e8-68187ac429f0-kube-api-access-dwsld\") pod \"must-gather-xtjbn\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.823407 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b0f0035-f0b8-48e7-86e8-68187ac429f0-must-gather-output\") pod \"must-gather-xtjbn\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.824007 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b0f0035-f0b8-48e7-86e8-68187ac429f0-must-gather-output\") pod \"must-gather-xtjbn\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.847053 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwsld\" (UniqueName: \"kubernetes.io/projected/6b0f0035-f0b8-48e7-86e8-68187ac429f0-kube-api-access-dwsld\") pod \"must-gather-xtjbn\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:32 crc kubenswrapper[4692]: I0309 10:02:32.996592 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:02:33 crc kubenswrapper[4692]: I0309 10:02:33.275920 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-t2242/must-gather-xtjbn"] Mar 09 10:02:33 crc kubenswrapper[4692]: I0309 10:02:33.729087 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t2242/must-gather-xtjbn" event={"ID":"6b0f0035-f0b8-48e7-86e8-68187ac429f0","Type":"ContainerStarted","Data":"6c97793d321884c4cd60ed0243d5a78f01f24fc28199874f3c99bf05cd00d83c"} Mar 09 10:02:40 crc kubenswrapper[4692]: I0309 10:02:40.918793 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t2242/must-gather-xtjbn" event={"ID":"6b0f0035-f0b8-48e7-86e8-68187ac429f0","Type":"ContainerStarted","Data":"67a68637f9be92f1d4a07c1d7c1b38576c485e2ae27e5c1a99c00c23d628d107"} Mar 09 10:02:41 crc kubenswrapper[4692]: I0309 10:02:41.932259 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t2242/must-gather-xtjbn" event={"ID":"6b0f0035-f0b8-48e7-86e8-68187ac429f0","Type":"ContainerStarted","Data":"ba26b4e27cf6d0cac52062c9020cae468498a112d80f52770fbc456cefe98679"} Mar 09 10:02:41 crc kubenswrapper[4692]: I0309 10:02:41.953973 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-t2242/must-gather-xtjbn" podStartSLOduration=2.754212377 podStartE2EDuration="9.953941437s" podCreationTimestamp="2026-03-09 10:02:32 +0000 UTC" firstStartedPulling="2026-03-09 10:02:33.27892072 +0000 UTC m=+2554.103656301" lastFinishedPulling="2026-03-09 10:02:40.47864978 +0000 UTC m=+2561.303385361" observedRunningTime="2026-03-09 10:02:41.951589751 +0000 UTC m=+2562.776325362" watchObservedRunningTime="2026-03-09 10:02:41.953941437 +0000 UTC m=+2562.778677018" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.165083 4692 scope.go:117] "RemoveContainer" containerID="75558d2113ea0b9d568a4ff68e63391d57278a629ac5b42eaa03b5eff4c8d930" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.204737 4692 scope.go:117] "RemoveContainer" containerID="1096f6b01f087e07b51c13627e1b1c3df945b3194a1b7183c99dcbc8326a04e4" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.248390 4692 scope.go:117] "RemoveContainer" containerID="14a05f337e0b5bd94fb5dd17b0c0e9bfc8a1552e5e92cd8b64220ba13a73e091" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.276879 4692 scope.go:117] "RemoveContainer" containerID="3f9de65df0f5efe5792cbd77ec85413b5a264133b28c837967efbdfbbfc787ba" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.306864 4692 scope.go:117] "RemoveContainer" containerID="e6a08f130ab3627ecaddd8d247938fc011b5419712a759cb11671b802d83efec" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.358192 4692 scope.go:117] "RemoveContainer" containerID="787409a72bfb1e6a5f6c151022b0437e7081acfbac56da9635bbd21aa799c6bb" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.396541 4692 scope.go:117] "RemoveContainer" containerID="bf9b14a26aa62bffccf4424e87193645f9a38e71e5f8b1f088d70b9bc4b030d6" Mar 09 10:03:07 crc kubenswrapper[4692]: I0309 10:03:07.433667 4692 scope.go:117] "RemoveContainer" containerID="ddcede9d7cd3b4f95753400e47bbe72d2e55b9a374e9f290f25d0a7aa7289b38" Mar 09 10:03:25 crc kubenswrapper[4692]: I0309 10:03:25.845618 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw_8901d82d-d48e-4b38-b746-78df5bf40e7d/util/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.058936 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw_8901d82d-d48e-4b38-b746-78df5bf40e7d/pull/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.072287 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw_8901d82d-d48e-4b38-b746-78df5bf40e7d/util/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.094723 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw_8901d82d-d48e-4b38-b746-78df5bf40e7d/pull/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.261341 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw_8901d82d-d48e-4b38-b746-78df5bf40e7d/pull/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.305464 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw_8901d82d-d48e-4b38-b746-78df5bf40e7d/extract/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.333332 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_30e5b271560c361eaa311e71b932ea1535005918a3724116e4108e6d0ar8vkw_8901d82d-d48e-4b38-b746-78df5bf40e7d/util/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.507900 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4_e6478415-5d9c-44fa-aed4-da919022fa54/util/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.704327 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4_e6478415-5d9c-44fa-aed4-da919022fa54/pull/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.744004 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4_e6478415-5d9c-44fa-aed4-da919022fa54/pull/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.746999 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4_e6478415-5d9c-44fa-aed4-da919022fa54/util/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.887759 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4_e6478415-5d9c-44fa-aed4-da919022fa54/util/0.log" Mar 09 10:03:26 crc kubenswrapper[4692]: I0309 10:03:26.922998 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4_e6478415-5d9c-44fa-aed4-da919022fa54/extract/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.158964 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590tgbz4_e6478415-5d9c-44fa-aed4-da919022fa54/pull/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.286607 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq_093b8260-b333-4f8e-a8b2-8313868609be/util/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.540662 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq_093b8260-b333-4f8e-a8b2-8313868609be/util/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.543428 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq_093b8260-b333-4f8e-a8b2-8313868609be/pull/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.581778 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq_093b8260-b333-4f8e-a8b2-8313868609be/pull/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.767387 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq_093b8260-b333-4f8e-a8b2-8313868609be/util/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.816754 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq_093b8260-b333-4f8e-a8b2-8313868609be/pull/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.831768 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40f94hq_093b8260-b333-4f8e-a8b2-8313868609be/extract/0.log" Mar 09 10:03:27 crc kubenswrapper[4692]: I0309 10:03:27.984446 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw_09cae4f4-3bef-442c-88ad-a84932acf202/util/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.200391 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw_09cae4f4-3bef-442c-88ad-a84932acf202/util/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.211992 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw_09cae4f4-3bef-442c-88ad-a84932acf202/pull/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.216250 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw_09cae4f4-3bef-442c-88ad-a84932acf202/pull/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.407241 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw_09cae4f4-3bef-442c-88ad-a84932acf202/util/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.436250 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw_09cae4f4-3bef-442c-88ad-a84932acf202/pull/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.442603 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da6tfskw_09cae4f4-3bef-442c-88ad-a84932acf202/extract/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.698474 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-index-tjnxd_4da32689-47a4-449d-af7b-5eedecfe8e01/registry-server/0.log" Mar 09 10:03:28 crc kubenswrapper[4692]: I0309 10:03:28.901543 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt_381ff66a-6f6e-49c5-9c29-d414b7200dae/util/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.119406 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt_381ff66a-6f6e-49c5-9c29-d414b7200dae/util/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.183741 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt_381ff66a-6f6e-49c5-9c29-d414b7200dae/pull/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.193357 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt_381ff66a-6f6e-49c5-9c29-d414b7200dae/pull/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.471776 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt_381ff66a-6f6e-49c5-9c29-d414b7200dae/util/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.474696 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt_381ff66a-6f6e-49c5-9c29-d414b7200dae/extract/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.515740 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_c5dd21a97976f00f72ca6d3429b14fbdb328dc08cfd92939eb7c42ec8c6jrwt_381ff66a-6f6e-49c5-9c29-d414b7200dae/pull/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.762153 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x_4f6faf73-b691-41b4-a0d7-f653bc0587c6/util/0.log" Mar 09 10:03:29 crc kubenswrapper[4692]: I0309 10:03:29.989601 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x_4f6faf73-b691-41b4-a0d7-f653bc0587c6/pull/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.040919 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x_4f6faf73-b691-41b4-a0d7-f653bc0587c6/pull/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.100384 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x_4f6faf73-b691-41b4-a0d7-f653bc0587c6/util/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.280269 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x_4f6faf73-b691-41b4-a0d7-f653bc0587c6/util/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.383511 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x_4f6faf73-b691-41b4-a0d7-f653bc0587c6/pull/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.393328 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_fa7dbe44ed02a3c936bbb063e46b2d7f493c0da97a741ca9b6b393031ebx59x_4f6faf73-b691-41b4-a0d7-f653bc0587c6/extract/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.569732 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5d6bc4dc48-7m6ss_ac73a80c-7cf2-44eb-b384-dec4e2b20783/manager/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.669241 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-ngw2q_a26e421d-abf7-477e-bad5-6ce43d8dd862/registry-server/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.878973 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5b7b95b7c4-9d5fb_d405d006-df42-4994-a407-0c3779dcf45c/manager/0.log" Mar 09 10:03:30 crc kubenswrapper[4692]: I0309 10:03:30.938553 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-682hs_2f8455e8-81a6-427c-ad1b-9b63aed74e4d/registry-server/0.log" Mar 09 10:03:31 crc kubenswrapper[4692]: I0309 10:03:31.213023 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-b4fb85cb4-f9kct_3336031e-a09e-471f-858c-f9c66f040451/manager/0.log" Mar 09 10:03:31 crc kubenswrapper[4692]: I0309 10:03:31.341044 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-54n7l_8733e667-ec14-4dbd-919d-c69774fdfec0/registry-server/0.log" Mar 09 10:03:31 crc kubenswrapper[4692]: I0309 10:03:31.484587 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-7pps2_335311a9-fa11-4fd5-b50e-29605ec7037e/operator/0.log" Mar 09 10:03:31 crc kubenswrapper[4692]: I0309 10:03:31.778212 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-4qdpt_eb637dd6-49f9-46a6-b3d8-c5e2f41921dc/registry-server/0.log" Mar 09 10:03:31 crc kubenswrapper[4692]: I0309 10:03:31.952259 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64d5945c66-k4v8x_9f91a180-3832-448a-9744-efdb956c9bd0/manager/0.log" Mar 09 10:03:31 crc kubenswrapper[4692]: I0309 10:03:31.992081 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6b7c9bbf4c-49qbg_d6362adf-592b-479c-ab1f-6dfd352e0570/manager/0.log" Mar 09 10:03:31 crc kubenswrapper[4692]: I0309 10:03:31.998109 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-2mmxr_31360b66-7736-41cf-bbff-ff901238bd28/registry-server/0.log" Mar 09 10:03:47 crc kubenswrapper[4692]: I0309 10:03:47.604814 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bng8j_14a9e924-9e6f-4a5e-82f4-ea2b52ae4334/control-plane-machine-set-operator/0.log" Mar 09 10:03:47 crc kubenswrapper[4692]: I0309 10:03:47.810929 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7gw55_33a84922-5581-4432-b6d9-d5a9edd7defe/kube-rbac-proxy/0.log" Mar 09 10:03:47 crc kubenswrapper[4692]: I0309 10:03:47.828369 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7gw55_33a84922-5581-4432-b6d9-d5a9edd7defe/machine-api-operator/0.log" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.142629 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550844-zbnlc"] Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.153383 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550844-zbnlc"] Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.153528 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550844-zbnlc" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.161432 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.161451 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.164461 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.291435 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdpzx\" (UniqueName: \"kubernetes.io/projected/2176b596-3849-486b-b77b-a75ea72fcb91-kube-api-access-kdpzx\") pod \"auto-csr-approver-29550844-zbnlc\" (UID: \"2176b596-3849-486b-b77b-a75ea72fcb91\") " pod="openshift-infra/auto-csr-approver-29550844-zbnlc" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.393539 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdpzx\" (UniqueName: \"kubernetes.io/projected/2176b596-3849-486b-b77b-a75ea72fcb91-kube-api-access-kdpzx\") pod \"auto-csr-approver-29550844-zbnlc\" (UID: \"2176b596-3849-486b-b77b-a75ea72fcb91\") " pod="openshift-infra/auto-csr-approver-29550844-zbnlc" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.413918 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdpzx\" (UniqueName: \"kubernetes.io/projected/2176b596-3849-486b-b77b-a75ea72fcb91-kube-api-access-kdpzx\") pod \"auto-csr-approver-29550844-zbnlc\" (UID: \"2176b596-3849-486b-b77b-a75ea72fcb91\") " pod="openshift-infra/auto-csr-approver-29550844-zbnlc" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.477212 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550844-zbnlc" Mar 09 10:04:00 crc kubenswrapper[4692]: I0309 10:04:00.955624 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550844-zbnlc"] Mar 09 10:04:01 crc kubenswrapper[4692]: I0309 10:04:01.720901 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550844-zbnlc" event={"ID":"2176b596-3849-486b-b77b-a75ea72fcb91","Type":"ContainerStarted","Data":"118de05c8fd452a64ae64704145ed61fb4c0f2cfbd2012c35c41d8ce897c6a97"} Mar 09 10:04:02 crc kubenswrapper[4692]: I0309 10:04:02.737619 4692 generic.go:334] "Generic (PLEG): container finished" podID="2176b596-3849-486b-b77b-a75ea72fcb91" containerID="e1f21daa3aa2d356e4a9f8c40e65600d183c00dcee7a525201cdbdab32afb9f9" exitCode=0 Mar 09 10:04:02 crc kubenswrapper[4692]: I0309 10:04:02.737714 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550844-zbnlc" event={"ID":"2176b596-3849-486b-b77b-a75ea72fcb91","Type":"ContainerDied","Data":"e1f21daa3aa2d356e4a9f8c40e65600d183c00dcee7a525201cdbdab32afb9f9"} Mar 09 10:04:04 crc kubenswrapper[4692]: I0309 10:04:04.078857 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550844-zbnlc" Mar 09 10:04:04 crc kubenswrapper[4692]: I0309 10:04:04.163026 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdpzx\" (UniqueName: \"kubernetes.io/projected/2176b596-3849-486b-b77b-a75ea72fcb91-kube-api-access-kdpzx\") pod \"2176b596-3849-486b-b77b-a75ea72fcb91\" (UID: \"2176b596-3849-486b-b77b-a75ea72fcb91\") " Mar 09 10:04:04 crc kubenswrapper[4692]: I0309 10:04:04.330391 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2176b596-3849-486b-b77b-a75ea72fcb91-kube-api-access-kdpzx" (OuterVolumeSpecName: "kube-api-access-kdpzx") pod "2176b596-3849-486b-b77b-a75ea72fcb91" (UID: "2176b596-3849-486b-b77b-a75ea72fcb91"). InnerVolumeSpecName "kube-api-access-kdpzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:04:04 crc kubenswrapper[4692]: I0309 10:04:04.367133 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdpzx\" (UniqueName: \"kubernetes.io/projected/2176b596-3849-486b-b77b-a75ea72fcb91-kube-api-access-kdpzx\") on node \"crc\" DevicePath \"\"" Mar 09 10:04:04 crc kubenswrapper[4692]: I0309 10:04:04.758097 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550844-zbnlc" event={"ID":"2176b596-3849-486b-b77b-a75ea72fcb91","Type":"ContainerDied","Data":"118de05c8fd452a64ae64704145ed61fb4c0f2cfbd2012c35c41d8ce897c6a97"} Mar 09 10:04:04 crc kubenswrapper[4692]: I0309 10:04:04.758142 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="118de05c8fd452a64ae64704145ed61fb4c0f2cfbd2012c35c41d8ce897c6a97" Mar 09 10:04:04 crc kubenswrapper[4692]: I0309 10:04:04.758221 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550844-zbnlc" Mar 09 10:04:05 crc kubenswrapper[4692]: I0309 10:04:05.155364 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550838-9zpjw"] Mar 09 10:04:05 crc kubenswrapper[4692]: I0309 10:04:05.160509 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550838-9zpjw"] Mar 09 10:04:06 crc kubenswrapper[4692]: I0309 10:04:06.082058 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="881ca87a-0f16-428d-b4e7-f9832b7b6f3e" path="/var/lib/kubelet/pods/881ca87a-0f16-428d-b4e7-f9832b7b6f3e/volumes" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.635641 4692 scope.go:117] "RemoveContainer" containerID="d6cc48a65007fc5e3a81661d7c4042baff65e5b33a69f27bcde40c180d1cea5e" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.666259 4692 scope.go:117] "RemoveContainer" containerID="44a1e7077d33ca5a2fb5c5b946c9afeb7a6332eeb50fc07e4b77cda6a08bfa99" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.747756 4692 scope.go:117] "RemoveContainer" containerID="934b15b36056a6e0d51982da677933fd90ed6ecdec82a15b2d71712f3a7a696f" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.814134 4692 scope.go:117] "RemoveContainer" containerID="a4b7b2479f08482c464c10d79b90425486c8c45a8b1d8fb433f36e0d310223c6" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.878528 4692 scope.go:117] "RemoveContainer" containerID="fff1aa1f2b5e7d05320624b5bef883ca44ed6de52835fb6652c1ea1c0dcfd75c" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.904075 4692 scope.go:117] "RemoveContainer" containerID="53b952ace5e9f05efb011c7e0e6bf396f45cc2ae4e531b8bd2bf38a1c7160c67" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.945263 4692 scope.go:117] "RemoveContainer" containerID="9e74e40ad1920b57ede3cf39989dae9ba4843dc6b52db3800f65d3b52cfe29b4" Mar 09 10:04:07 crc kubenswrapper[4692]: I0309 10:04:07.975589 4692 scope.go:117] "RemoveContainer" containerID="7dfaaa7954a17d8a987b64ba7c2dcec0ad6c6676b10c7fbefb94f71b6a62d9a8" Mar 09 10:04:20 crc kubenswrapper[4692]: I0309 10:04:20.354976 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-fhbtn_48590eda-9d17-4bed-9911-4aad14e46289/kube-rbac-proxy/0.log" Mar 09 10:04:20 crc kubenswrapper[4692]: I0309 10:04:20.385770 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-fhbtn_48590eda-9d17-4bed-9911-4aad14e46289/controller/0.log" Mar 09 10:04:20 crc kubenswrapper[4692]: I0309 10:04:20.556057 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-frr-files/0.log" Mar 09 10:04:20 crc kubenswrapper[4692]: I0309 10:04:20.791858 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-reloader/0.log" Mar 09 10:04:20 crc kubenswrapper[4692]: I0309 10:04:20.793484 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-metrics/0.log" Mar 09 10:04:20 crc kubenswrapper[4692]: I0309 10:04:20.834171 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-reloader/0.log" Mar 09 10:04:20 crc kubenswrapper[4692]: I0309 10:04:20.868298 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-frr-files/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.066771 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-reloader/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.096357 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-frr-files/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.143760 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-metrics/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.148095 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-metrics/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.361428 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-metrics/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.362489 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-reloader/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.387960 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/controller/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.429968 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/cp-frr-files/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.593604 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/frr-metrics/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.655314 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/kube-rbac-proxy/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.672118 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/kube-rbac-proxy-frr/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.860751 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/reloader/0.log" Mar 09 10:04:21 crc kubenswrapper[4692]: I0309 10:04:21.910761 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-n6cmq_c1e68ce3-7af3-4509-a31e-81f42f223aa9/frr-k8s-webhook-server/0.log" Mar 09 10:04:22 crc kubenswrapper[4692]: I0309 10:04:22.186695 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-566dcbb64d-j485l_aeb7fbbf-aa00-4666-aa7f-55abaac713c4/manager/0.log" Mar 09 10:04:22 crc kubenswrapper[4692]: I0309 10:04:22.525521 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5d76797467-xnptd_a9359a7b-47b3-4747-aa40-2633db5e907a/webhook-server/0.log" Mar 09 10:04:22 crc kubenswrapper[4692]: I0309 10:04:22.666864 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6nhdr_8b748f8c-3824-4228-bec4-634e2ed9f324/kube-rbac-proxy/0.log" Mar 09 10:04:22 crc kubenswrapper[4692]: I0309 10:04:22.957222 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6nhdr_8b748f8c-3824-4228-bec4-634e2ed9f324/speaker/0.log" Mar 09 10:04:23 crc kubenswrapper[4692]: I0309 10:04:23.299650 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-5b8n8_32f4683a-ccb2-44de-a220-6e279f138d99/frr/0.log" Mar 09 10:04:38 crc kubenswrapper[4692]: I0309 10:04:38.825915 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-api-5b4dcd568-j4snq_47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4/barbican-api/0.log" Mar 09 10:04:38 crc kubenswrapper[4692]: I0309 10:04:38.899030 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-api-5b4dcd568-j4snq_47dcccf0-cea4-48ce-91cc-dc02ddeb1ab4/barbican-api-log/0.log" Mar 09 10:04:39 crc kubenswrapper[4692]: I0309 10:04:39.059245 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-db-sync-b22vl_189d4c9b-7202-4625-a552-fa66691a4bb7/barbican-db-sync/0.log" Mar 09 10:04:39 crc kubenswrapper[4692]: I0309 10:04:39.124883 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-keystone-listener-74f57bcf5d-nbr2p_e8de0018-50da-4b21-a94c-ef673de69ab1/barbican-keystone-listener/0.log" Mar 09 10:04:39 crc kubenswrapper[4692]: I0309 10:04:39.317691 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-keystone-listener-74f57bcf5d-nbr2p_e8de0018-50da-4b21-a94c-ef673de69ab1/barbican-keystone-listener-log/0.log" Mar 09 10:04:39 crc kubenswrapper[4692]: I0309 10:04:39.331914 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-worker-786f5f954f-mmjsx_2a94f29a-c39a-4c48-be28-472570beb2b3/barbican-worker/0.log" Mar 09 10:04:39 crc kubenswrapper[4692]: I0309 10:04:39.332818 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_barbican-worker-786f5f954f-mmjsx_2a94f29a-c39a-4c48-be28-472570beb2b3/barbican-worker-log/0.log" Mar 09 10:04:39 crc kubenswrapper[4692]: I0309 10:04:39.538225 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_keystone-cron-29550841-psdz4_e4861760-6b7f-4777-a35d-bdb3614c4d39/keystone-cron/0.log" Mar 09 10:04:39 crc kubenswrapper[4692]: I0309 10:04:39.818358 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_keystone-88b4f4fb5-8bz88_33b9dc03-c993-4bcd-9ccc-6ef1a6e7dd36/keystone-api/0.log" Mar 09 10:04:40 crc kubenswrapper[4692]: I0309 10:04:40.100313 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-0_7b194e65-c0c6-4ab1-966d-99ce335b719b/mysql-bootstrap/0.log" Mar 09 10:04:40 crc kubenswrapper[4692]: I0309 10:04:40.290520 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-0_7b194e65-c0c6-4ab1-966d-99ce335b719b/mysql-bootstrap/0.log" Mar 09 10:04:40 crc kubenswrapper[4692]: I0309 10:04:40.411621 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-0_7b194e65-c0c6-4ab1-966d-99ce335b719b/galera/0.log" Mar 09 10:04:40 crc kubenswrapper[4692]: I0309 10:04:40.550975 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-1_75f90081-7951-418d-9c49-5aa141ee85b0/mysql-bootstrap/0.log" Mar 09 10:04:40 crc kubenswrapper[4692]: I0309 10:04:40.828264 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-1_75f90081-7951-418d-9c49-5aa141ee85b0/galera/0.log" Mar 09 10:04:40 crc kubenswrapper[4692]: I0309 10:04:40.845666 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-1_75f90081-7951-418d-9c49-5aa141ee85b0/mysql-bootstrap/0.log" Mar 09 10:04:41 crc kubenswrapper[4692]: I0309 10:04:41.238819 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-2_87a076b1-43df-4290-903a-abc067ba6cad/mysql-bootstrap/0.log" Mar 09 10:04:41 crc kubenswrapper[4692]: I0309 10:04:41.466731 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-2_87a076b1-43df-4290-903a-abc067ba6cad/mysql-bootstrap/0.log" Mar 09 10:04:41 crc kubenswrapper[4692]: I0309 10:04:41.558023 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_openstack-galera-2_87a076b1-43df-4290-903a-abc067ba6cad/galera/0.log" Mar 09 10:04:41 crc kubenswrapper[4692]: I0309 10:04:41.748356 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_rabbitmq-server-0_8075e838-fcea-4974-aead-79a1c9f94aa9/setup-container/0.log" Mar 09 10:04:41 crc kubenswrapper[4692]: I0309 10:04:41.874217 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_rabbitmq-server-0_8075e838-fcea-4974-aead-79a1c9f94aa9/setup-container/0.log" Mar 09 10:04:41 crc kubenswrapper[4692]: I0309 10:04:41.898185 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_rabbitmq-server-0_8075e838-fcea-4974-aead-79a1c9f94aa9/rabbitmq/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.116535 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-proxy-76c998454c-69ksf_1028ab76-0253-4e6d-9c68-d0b4758a9182/proxy-server/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.303866 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-proxy-76c998454c-69ksf_1028ab76-0253-4e6d-9c68-d0b4758a9182/proxy-httpd/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.332935 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-ring-rebalance-p6vr9_5a69287f-b725-45f6-85ed-9ce460d1e031/swift-ring-rebalance/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.363775 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_memcached-0_67894e62-3a7a-43a0-8497-74c261fbe1bb/memcached/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.520576 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/account-reaper/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.545127 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/account-auditor/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.591701 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/account-replicator/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.714527 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/account-server/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.755677 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/container-auditor/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.757520 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/container-replicator/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.812750 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/container-server/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.971486 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/object-expirer/0.log" Mar 09 10:04:42 crc kubenswrapper[4692]: I0309 10:04:42.973094 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/object-auditor/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.002364 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/container-updater/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.061751 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/object-replicator/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.167591 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/object-server/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.205106 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/object-updater/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.271364 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/rsync/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.283305 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-0_5b3f7c6f-e373-484e-bbe1-1e7b14478b9f/swift-recon-cron/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.398684 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/account-auditor/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.441941 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/account-reaper/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.504078 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/account-replicator/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.505387 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/account-server/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.569617 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/container-auditor/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.673097 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/container-replicator/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.682838 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/container-server/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.754646 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/container-updater/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.777803 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/object-auditor/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.793043 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/object-expirer/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.890247 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/object-replicator/0.log" Mar 09 10:04:43 crc kubenswrapper[4692]: I0309 10:04:43.946002 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/object-server/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.019117 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/object-updater/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.045957 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/swift-recon-cron/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.070019 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-1_e35fd62c-2bc9-44a1-9541-691741fb1d69/rsync/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.198324 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/account-auditor/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.255331 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/account-reaper/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.320771 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/account-server/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.323701 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/account-replicator/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.407581 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/container-auditor/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.478058 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/container-server/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.500216 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/container-replicator/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.542923 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/object-auditor/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.546227 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/container-updater/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.628140 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/object-expirer/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.673825 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/object-replicator/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.734327 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/object-server/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.754028 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/rsync/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.771766 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/object-updater/0.log" Mar 09 10:04:44 crc kubenswrapper[4692]: I0309 10:04:44.844983 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/swift-kuttl-tests_swift-storage-2_367e8bc6-334a-490d-a1a5-4969f2a74a52/swift-recon-cron/0.log" Mar 09 10:04:47 crc kubenswrapper[4692]: I0309 10:04:47.624351 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 10:04:47 crc kubenswrapper[4692]: I0309 10:04:47.625217 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 10:04:59 crc kubenswrapper[4692]: I0309 10:04:59.486691 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gfw8w_8065edd5-e74a-439a-8e0a-e198d4915738/extract-utilities/0.log" Mar 09 10:04:59 crc kubenswrapper[4692]: I0309 10:04:59.674141 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gfw8w_8065edd5-e74a-439a-8e0a-e198d4915738/extract-content/0.log" Mar 09 10:04:59 crc kubenswrapper[4692]: I0309 10:04:59.681132 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gfw8w_8065edd5-e74a-439a-8e0a-e198d4915738/extract-utilities/0.log" Mar 09 10:04:59 crc kubenswrapper[4692]: I0309 10:04:59.690184 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gfw8w_8065edd5-e74a-439a-8e0a-e198d4915738/extract-content/0.log" Mar 09 10:04:59 crc kubenswrapper[4692]: I0309 10:04:59.932796 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gfw8w_8065edd5-e74a-439a-8e0a-e198d4915738/extract-utilities/0.log" Mar 09 10:04:59 crc kubenswrapper[4692]: I0309 10:04:59.951515 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gfw8w_8065edd5-e74a-439a-8e0a-e198d4915738/extract-content/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.196438 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wp4gf_2c2af711-e30f-4153-85b8-864a5250e4e2/extract-utilities/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.497211 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wp4gf_2c2af711-e30f-4153-85b8-864a5250e4e2/extract-content/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.561110 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gfw8w_8065edd5-e74a-439a-8e0a-e198d4915738/registry-server/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.567905 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wp4gf_2c2af711-e30f-4153-85b8-864a5250e4e2/extract-content/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.581533 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wp4gf_2c2af711-e30f-4153-85b8-864a5250e4e2/extract-utilities/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.737125 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wp4gf_2c2af711-e30f-4153-85b8-864a5250e4e2/extract-utilities/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.762791 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wp4gf_2c2af711-e30f-4153-85b8-864a5250e4e2/extract-content/0.log" Mar 09 10:05:00 crc kubenswrapper[4692]: I0309 10:05:00.960728 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg_e6640a53-9679-4b36-a32d-c0fe67ecfe26/util/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.197919 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wp4gf_2c2af711-e30f-4153-85b8-864a5250e4e2/registry-server/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.305455 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg_e6640a53-9679-4b36-a32d-c0fe67ecfe26/util/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.313780 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg_e6640a53-9679-4b36-a32d-c0fe67ecfe26/pull/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.317241 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg_e6640a53-9679-4b36-a32d-c0fe67ecfe26/pull/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.687841 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg_e6640a53-9679-4b36-a32d-c0fe67ecfe26/util/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.697603 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg_e6640a53-9679-4b36-a32d-c0fe67ecfe26/extract/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.732531 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4hvdkg_e6640a53-9679-4b36-a32d-c0fe67ecfe26/pull/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.915582 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wvcsq_74fc14ae-20b4-429a-8f94-886a6bc85d3e/marketplace-operator/0.log" Mar 09 10:05:01 crc kubenswrapper[4692]: I0309 10:05:01.973118 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pjjn2_2be7a257-3244-484e-9747-08d81edfd021/extract-utilities/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.104508 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pjjn2_2be7a257-3244-484e-9747-08d81edfd021/extract-content/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.132883 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pjjn2_2be7a257-3244-484e-9747-08d81edfd021/extract-utilities/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.157721 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pjjn2_2be7a257-3244-484e-9747-08d81edfd021/extract-content/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.347842 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pjjn2_2be7a257-3244-484e-9747-08d81edfd021/extract-content/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.348969 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pjjn2_2be7a257-3244-484e-9747-08d81edfd021/extract-utilities/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.455271 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pjjn2_2be7a257-3244-484e-9747-08d81edfd021/registry-server/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.594902 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6vw9_335a9dc7-133c-4667-bb5f-3ba7598aeefc/extract-utilities/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.799544 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6vw9_335a9dc7-133c-4667-bb5f-3ba7598aeefc/extract-utilities/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.806320 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6vw9_335a9dc7-133c-4667-bb5f-3ba7598aeefc/extract-content/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.812798 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6vw9_335a9dc7-133c-4667-bb5f-3ba7598aeefc/extract-content/0.log" Mar 09 10:05:02 crc kubenswrapper[4692]: I0309 10:05:02.987172 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6vw9_335a9dc7-133c-4667-bb5f-3ba7598aeefc/extract-utilities/0.log" Mar 09 10:05:03 crc kubenswrapper[4692]: I0309 10:05:03.005434 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6vw9_335a9dc7-133c-4667-bb5f-3ba7598aeefc/extract-content/0.log" Mar 09 10:05:03 crc kubenswrapper[4692]: I0309 10:05:03.564210 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-n6vw9_335a9dc7-133c-4667-bb5f-3ba7598aeefc/registry-server/0.log" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.162376 4692 scope.go:117] "RemoveContainer" containerID="c39a6ec6b5787e28d96f0698d28fc4e4749cef8d8f7381f41cb412e99181770e" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.205881 4692 scope.go:117] "RemoveContainer" containerID="cd7921286faa7a351505044018ae816ea4516eafda0caaaf0762e9442f8f5bf6" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.238553 4692 scope.go:117] "RemoveContainer" containerID="49c59ddc3778b11939e3d214d4ca8519b433ac2550c237dcc16a09d5e064ff7a" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.272978 4692 scope.go:117] "RemoveContainer" containerID="ebcc8354c1009007af26c0d9c3d6d8574fb0427f871aa435511b9319b399e4f4" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.307006 4692 scope.go:117] "RemoveContainer" containerID="d4d672cf55b6bbc477e4745fbea789482e5f0eb874d6fe2f0acee870e9211993" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.353096 4692 scope.go:117] "RemoveContainer" containerID="589cc4f0a066ba0c299651050033de51e68e2be669e97613b888ff4b74702a46" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.390031 4692 scope.go:117] "RemoveContainer" containerID="9fcd1d22bd9e043bfb4a986f66b936f27793554af5d6483b965bfff394a1e29d" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.423104 4692 scope.go:117] "RemoveContainer" containerID="688051e31d1d9865ebe410f5447f2518d7bd92a0296c388ba4d03d8073eb57ab" Mar 09 10:05:08 crc kubenswrapper[4692]: I0309 10:05:08.453476 4692 scope.go:117] "RemoveContainer" containerID="97f4b3321a2f2c55b07b999e3c38bc44f9dd9bb5923fd87a74d7c0245b9bf2db" Mar 09 10:05:17 crc kubenswrapper[4692]: I0309 10:05:17.623816 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 10:05:17 crc kubenswrapper[4692]: I0309 10:05:17.624700 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.624833 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.625733 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.625826 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.627482 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c4aca92fab9686fc64015cb64a831b1e43d53698f0ab8c3789ce9a9e0a536fa"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.627619 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://0c4aca92fab9686fc64015cb64a831b1e43d53698f0ab8c3789ce9a9e0a536fa" gracePeriod=600 Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.768495 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="0c4aca92fab9686fc64015cb64a831b1e43d53698f0ab8c3789ce9a9e0a536fa" exitCode=0 Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.768533 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"0c4aca92fab9686fc64015cb64a831b1e43d53698f0ab8c3789ce9a9e0a536fa"} Mar 09 10:05:47 crc kubenswrapper[4692]: I0309 10:05:47.768950 4692 scope.go:117] "RemoveContainer" containerID="d2e15eb825a63c2ac3bf39fc57d1568601a32b592d49bcadf7dac7718d4381ff" Mar 09 10:05:48 crc kubenswrapper[4692]: I0309 10:05:48.781510 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerStarted","Data":"56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb"} Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.170949 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550846-64vgt"] Mar 09 10:06:00 crc kubenswrapper[4692]: E0309 10:06:00.172297 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2176b596-3849-486b-b77b-a75ea72fcb91" containerName="oc" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.172316 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="2176b596-3849-486b-b77b-a75ea72fcb91" containerName="oc" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.172500 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="2176b596-3849-486b-b77b-a75ea72fcb91" containerName="oc" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.173173 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550846-64vgt" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.178322 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.179402 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550846-64vgt"] Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.180212 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.181336 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.301521 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-675sc\" (UniqueName: \"kubernetes.io/projected/1066f819-6741-40be-83fa-11ef33265fe5-kube-api-access-675sc\") pod \"auto-csr-approver-29550846-64vgt\" (UID: \"1066f819-6741-40be-83fa-11ef33265fe5\") " pod="openshift-infra/auto-csr-approver-29550846-64vgt" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.403850 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-675sc\" (UniqueName: \"kubernetes.io/projected/1066f819-6741-40be-83fa-11ef33265fe5-kube-api-access-675sc\") pod \"auto-csr-approver-29550846-64vgt\" (UID: \"1066f819-6741-40be-83fa-11ef33265fe5\") " pod="openshift-infra/auto-csr-approver-29550846-64vgt" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.427119 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-675sc\" (UniqueName: \"kubernetes.io/projected/1066f819-6741-40be-83fa-11ef33265fe5-kube-api-access-675sc\") pod \"auto-csr-approver-29550846-64vgt\" (UID: \"1066f819-6741-40be-83fa-11ef33265fe5\") " pod="openshift-infra/auto-csr-approver-29550846-64vgt" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.501626 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550846-64vgt" Mar 09 10:06:00 crc kubenswrapper[4692]: I0309 10:06:00.953701 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550846-64vgt"] Mar 09 10:06:01 crc kubenswrapper[4692]: I0309 10:06:01.908312 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550846-64vgt" event={"ID":"1066f819-6741-40be-83fa-11ef33265fe5","Type":"ContainerStarted","Data":"93483cca843fefeca847a0cb34417f9076c8124be90a6da81a500c0cb5578f5f"} Mar 09 10:06:02 crc kubenswrapper[4692]: I0309 10:06:02.919252 4692 generic.go:334] "Generic (PLEG): container finished" podID="1066f819-6741-40be-83fa-11ef33265fe5" containerID="8e98b816fafbff5f3888e11142a47bfedd9970c5a8b9efa4d671d76f851dfd83" exitCode=0 Mar 09 10:06:02 crc kubenswrapper[4692]: I0309 10:06:02.919357 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550846-64vgt" event={"ID":"1066f819-6741-40be-83fa-11ef33265fe5","Type":"ContainerDied","Data":"8e98b816fafbff5f3888e11142a47bfedd9970c5a8b9efa4d671d76f851dfd83"} Mar 09 10:06:04 crc kubenswrapper[4692]: I0309 10:06:04.251361 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550846-64vgt" Mar 09 10:06:04 crc kubenswrapper[4692]: I0309 10:06:04.371985 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-675sc\" (UniqueName: \"kubernetes.io/projected/1066f819-6741-40be-83fa-11ef33265fe5-kube-api-access-675sc\") pod \"1066f819-6741-40be-83fa-11ef33265fe5\" (UID: \"1066f819-6741-40be-83fa-11ef33265fe5\") " Mar 09 10:06:04 crc kubenswrapper[4692]: I0309 10:06:04.379964 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1066f819-6741-40be-83fa-11ef33265fe5-kube-api-access-675sc" (OuterVolumeSpecName: "kube-api-access-675sc") pod "1066f819-6741-40be-83fa-11ef33265fe5" (UID: "1066f819-6741-40be-83fa-11ef33265fe5"). InnerVolumeSpecName "kube-api-access-675sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:06:04 crc kubenswrapper[4692]: I0309 10:06:04.475021 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-675sc\" (UniqueName: \"kubernetes.io/projected/1066f819-6741-40be-83fa-11ef33265fe5-kube-api-access-675sc\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:04 crc kubenswrapper[4692]: I0309 10:06:04.942274 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550846-64vgt" event={"ID":"1066f819-6741-40be-83fa-11ef33265fe5","Type":"ContainerDied","Data":"93483cca843fefeca847a0cb34417f9076c8124be90a6da81a500c0cb5578f5f"} Mar 09 10:06:04 crc kubenswrapper[4692]: I0309 10:06:04.942568 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93483cca843fefeca847a0cb34417f9076c8124be90a6da81a500c0cb5578f5f" Mar 09 10:06:04 crc kubenswrapper[4692]: I0309 10:06:04.942360 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550846-64vgt" Mar 09 10:06:05 crc kubenswrapper[4692]: I0309 10:06:05.322178 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550840-tqnz9"] Mar 09 10:06:05 crc kubenswrapper[4692]: I0309 10:06:05.329195 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550840-tqnz9"] Mar 09 10:06:06 crc kubenswrapper[4692]: I0309 10:06:06.082752 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef13db35-bc96-4421-93a0-e8690b87d300" path="/var/lib/kubelet/pods/ef13db35-bc96-4421-93a0-e8690b87d300/volumes" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.633141 4692 scope.go:117] "RemoveContainer" containerID="b6bf8462ad09bac39f0c708f34d236c6cb983e2b531ebe205c4055b651c8609d" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.672859 4692 scope.go:117] "RemoveContainer" containerID="b3e83c2f9748e85094f461c07e8e14fe2f5a9c383f06cab0ff4f24a491577c46" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.709067 4692 scope.go:117] "RemoveContainer" containerID="b286711ebf13e74405f3c83ef50f29eb75fda7aaded659743ddae1074d870598" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.766670 4692 scope.go:117] "RemoveContainer" containerID="8b57adaff31d14e0799698f3f2242fbd193f96cca51f751415e62c82941bfd4c" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.795061 4692 scope.go:117] "RemoveContainer" containerID="28a0514e56ea08d97f779af95ccabbff6fccf035fa321903c56314dbceb835a3" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.850323 4692 scope.go:117] "RemoveContainer" containerID="1ec9bcafe9f6d5c5553ec0dedfb27d9f211f302064fd33bfe663880e47c78589" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.889121 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gmttz"] Mar 09 10:06:08 crc kubenswrapper[4692]: E0309 10:06:08.889684 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1066f819-6741-40be-83fa-11ef33265fe5" containerName="oc" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.889713 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1066f819-6741-40be-83fa-11ef33265fe5" containerName="oc" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.889973 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1066f819-6741-40be-83fa-11ef33265fe5" containerName="oc" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.891360 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.896651 4692 scope.go:117] "RemoveContainer" containerID="302c063ddd7a1cd0bbb5da39cec9cc0d61214b8f1872ffad288902a0341b4d1c" Mar 09 10:06:08 crc kubenswrapper[4692]: I0309 10:06:08.902651 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmttz"] Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.067347 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-catalog-content\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.067395 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9tpd\" (UniqueName: \"kubernetes.io/projected/434276c5-b038-4f0d-a7f9-72471e3ba329-kube-api-access-c9tpd\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.067430 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-utilities\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.168604 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-catalog-content\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.168670 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9tpd\" (UniqueName: \"kubernetes.io/projected/434276c5-b038-4f0d-a7f9-72471e3ba329-kube-api-access-c9tpd\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.168703 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-utilities\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.169308 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-utilities\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.169549 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-catalog-content\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.196958 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9tpd\" (UniqueName: \"kubernetes.io/projected/434276c5-b038-4f0d-a7f9-72471e3ba329-kube-api-access-c9tpd\") pod \"community-operators-gmttz\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.243499 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:09 crc kubenswrapper[4692]: I0309 10:06:09.717903 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gmttz"] Mar 09 10:06:10 crc kubenswrapper[4692]: I0309 10:06:10.119289 4692 generic.go:334] "Generic (PLEG): container finished" podID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerID="c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a" exitCode=0 Mar 09 10:06:10 crc kubenswrapper[4692]: I0309 10:06:10.119426 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmttz" event={"ID":"434276c5-b038-4f0d-a7f9-72471e3ba329","Type":"ContainerDied","Data":"c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a"} Mar 09 10:06:10 crc kubenswrapper[4692]: I0309 10:06:10.120789 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmttz" event={"ID":"434276c5-b038-4f0d-a7f9-72471e3ba329","Type":"ContainerStarted","Data":"4fceb479370e9ed455db17ff52e6830114191bab1227b0787145886b3f057ed7"} Mar 09 10:06:11 crc kubenswrapper[4692]: I0309 10:06:11.134128 4692 generic.go:334] "Generic (PLEG): container finished" podID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerID="212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015" exitCode=0 Mar 09 10:06:11 crc kubenswrapper[4692]: I0309 10:06:11.134338 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmttz" event={"ID":"434276c5-b038-4f0d-a7f9-72471e3ba329","Type":"ContainerDied","Data":"212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015"} Mar 09 10:06:12 crc kubenswrapper[4692]: I0309 10:06:12.148591 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmttz" event={"ID":"434276c5-b038-4f0d-a7f9-72471e3ba329","Type":"ContainerStarted","Data":"8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31"} Mar 09 10:06:12 crc kubenswrapper[4692]: I0309 10:06:12.175939 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gmttz" podStartSLOduration=2.66969719 podStartE2EDuration="4.175911549s" podCreationTimestamp="2026-03-09 10:06:08 +0000 UTC" firstStartedPulling="2026-03-09 10:06:10.12169776 +0000 UTC m=+2770.946433341" lastFinishedPulling="2026-03-09 10:06:11.627912119 +0000 UTC m=+2772.452647700" observedRunningTime="2026-03-09 10:06:12.170240549 +0000 UTC m=+2772.994976140" watchObservedRunningTime="2026-03-09 10:06:12.175911549 +0000 UTC m=+2773.000647140" Mar 09 10:06:19 crc kubenswrapper[4692]: I0309 10:06:19.243897 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:19 crc kubenswrapper[4692]: I0309 10:06:19.244749 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:19 crc kubenswrapper[4692]: I0309 10:06:19.303430 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:20 crc kubenswrapper[4692]: I0309 10:06:20.276720 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:20 crc kubenswrapper[4692]: I0309 10:06:20.337407 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gmttz"] Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.257493 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gmttz" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="registry-server" containerID="cri-o://8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31" gracePeriod=2 Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.705610 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.839594 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-catalog-content\") pod \"434276c5-b038-4f0d-a7f9-72471e3ba329\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.839678 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9tpd\" (UniqueName: \"kubernetes.io/projected/434276c5-b038-4f0d-a7f9-72471e3ba329-kube-api-access-c9tpd\") pod \"434276c5-b038-4f0d-a7f9-72471e3ba329\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.839867 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-utilities\") pod \"434276c5-b038-4f0d-a7f9-72471e3ba329\" (UID: \"434276c5-b038-4f0d-a7f9-72471e3ba329\") " Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.841236 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-utilities" (OuterVolumeSpecName: "utilities") pod "434276c5-b038-4f0d-a7f9-72471e3ba329" (UID: "434276c5-b038-4f0d-a7f9-72471e3ba329"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.855559 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434276c5-b038-4f0d-a7f9-72471e3ba329-kube-api-access-c9tpd" (OuterVolumeSpecName: "kube-api-access-c9tpd") pod "434276c5-b038-4f0d-a7f9-72471e3ba329" (UID: "434276c5-b038-4f0d-a7f9-72471e3ba329"). InnerVolumeSpecName "kube-api-access-c9tpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.892528 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "434276c5-b038-4f0d-a7f9-72471e3ba329" (UID: "434276c5-b038-4f0d-a7f9-72471e3ba329"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.941618 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.941686 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/434276c5-b038-4f0d-a7f9-72471e3ba329-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:22 crc kubenswrapper[4692]: I0309 10:06:22.941697 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9tpd\" (UniqueName: \"kubernetes.io/projected/434276c5-b038-4f0d-a7f9-72471e3ba329-kube-api-access-c9tpd\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.267777 4692 generic.go:334] "Generic (PLEG): container finished" podID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerID="8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31" exitCode=0 Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.267838 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gmttz" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.267885 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmttz" event={"ID":"434276c5-b038-4f0d-a7f9-72471e3ba329","Type":"ContainerDied","Data":"8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31"} Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.268303 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gmttz" event={"ID":"434276c5-b038-4f0d-a7f9-72471e3ba329","Type":"ContainerDied","Data":"4fceb479370e9ed455db17ff52e6830114191bab1227b0787145886b3f057ed7"} Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.268362 4692 scope.go:117] "RemoveContainer" containerID="8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.311973 4692 scope.go:117] "RemoveContainer" containerID="212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.314788 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gmttz"] Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.339225 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gmttz"] Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.363037 4692 scope.go:117] "RemoveContainer" containerID="c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.383432 4692 scope.go:117] "RemoveContainer" containerID="8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31" Mar 09 10:06:23 crc kubenswrapper[4692]: E0309 10:06:23.384153 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31\": container with ID starting with 8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31 not found: ID does not exist" containerID="8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.384253 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31"} err="failed to get container status \"8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31\": rpc error: code = NotFound desc = could not find container \"8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31\": container with ID starting with 8bf5c8cbaa75cd2b09cc91eaf321a0ea96fa18b1d3360c12f578cb3b741efc31 not found: ID does not exist" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.384308 4692 scope.go:117] "RemoveContainer" containerID="212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015" Mar 09 10:06:23 crc kubenswrapper[4692]: E0309 10:06:23.384694 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015\": container with ID starting with 212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015 not found: ID does not exist" containerID="212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.384734 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015"} err="failed to get container status \"212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015\": rpc error: code = NotFound desc = could not find container \"212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015\": container with ID starting with 212a00400bb88e406814fb9a1571e7cdc0fd936f3a43bb33680f860b23144015 not found: ID does not exist" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.384761 4692 scope.go:117] "RemoveContainer" containerID="c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a" Mar 09 10:06:23 crc kubenswrapper[4692]: E0309 10:06:23.385042 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a\": container with ID starting with c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a not found: ID does not exist" containerID="c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a" Mar 09 10:06:23 crc kubenswrapper[4692]: I0309 10:06:23.385074 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a"} err="failed to get container status \"c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a\": rpc error: code = NotFound desc = could not find container \"c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a\": container with ID starting with c9c4264bbee45a89924b2ccfb4960630b69c10ff307cfe8d90b03798d0f2948a not found: ID does not exist" Mar 09 10:06:24 crc kubenswrapper[4692]: I0309 10:06:24.083471 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" path="/var/lib/kubelet/pods/434276c5-b038-4f0d-a7f9-72471e3ba329/volumes" Mar 09 10:06:29 crc kubenswrapper[4692]: I0309 10:06:29.335474 4692 generic.go:334] "Generic (PLEG): container finished" podID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerID="67a68637f9be92f1d4a07c1d7c1b38576c485e2ae27e5c1a99c00c23d628d107" exitCode=0 Mar 09 10:06:29 crc kubenswrapper[4692]: I0309 10:06:29.336102 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-t2242/must-gather-xtjbn" event={"ID":"6b0f0035-f0b8-48e7-86e8-68187ac429f0","Type":"ContainerDied","Data":"67a68637f9be92f1d4a07c1d7c1b38576c485e2ae27e5c1a99c00c23d628d107"} Mar 09 10:06:29 crc kubenswrapper[4692]: I0309 10:06:29.339330 4692 scope.go:117] "RemoveContainer" containerID="67a68637f9be92f1d4a07c1d7c1b38576c485e2ae27e5c1a99c00c23d628d107" Mar 09 10:06:29 crc kubenswrapper[4692]: I0309 10:06:29.992294 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t2242_must-gather-xtjbn_6b0f0035-f0b8-48e7-86e8-68187ac429f0/gather/0.log" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.502503 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ck7"] Mar 09 10:06:32 crc kubenswrapper[4692]: E0309 10:06:32.502916 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="extract-content" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.502932 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="extract-content" Mar 09 10:06:32 crc kubenswrapper[4692]: E0309 10:06:32.502955 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="registry-server" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.502962 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="registry-server" Mar 09 10:06:32 crc kubenswrapper[4692]: E0309 10:06:32.502979 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="extract-utilities" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.502985 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="extract-utilities" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.503174 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="434276c5-b038-4f0d-a7f9-72471e3ba329" containerName="registry-server" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.504360 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.519080 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ck7"] Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.610512 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-utilities\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.610720 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lfb8\" (UniqueName: \"kubernetes.io/projected/1902263a-5295-4bd6-99fa-c486f77b18ba-kube-api-access-5lfb8\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.610785 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-catalog-content\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.714049 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-utilities\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.714570 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lfb8\" (UniqueName: \"kubernetes.io/projected/1902263a-5295-4bd6-99fa-c486f77b18ba-kube-api-access-5lfb8\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.714602 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-catalog-content\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.715123 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-catalog-content\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.715296 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-utilities\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.749455 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lfb8\" (UniqueName: \"kubernetes.io/projected/1902263a-5295-4bd6-99fa-c486f77b18ba-kube-api-access-5lfb8\") pod \"redhat-marketplace-h4ck7\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:32 crc kubenswrapper[4692]: I0309 10:06:32.871582 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:33 crc kubenswrapper[4692]: I0309 10:06:33.325261 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ck7"] Mar 09 10:06:33 crc kubenswrapper[4692]: I0309 10:06:33.385929 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ck7" event={"ID":"1902263a-5295-4bd6-99fa-c486f77b18ba","Type":"ContainerStarted","Data":"665286daee24548b9307ca9824dfdb00731fa02ba1b1a181ea7b8201d7f60a69"} Mar 09 10:06:34 crc kubenswrapper[4692]: I0309 10:06:34.397902 4692 generic.go:334] "Generic (PLEG): container finished" podID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerID="c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c" exitCode=0 Mar 09 10:06:34 crc kubenswrapper[4692]: I0309 10:06:34.398035 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ck7" event={"ID":"1902263a-5295-4bd6-99fa-c486f77b18ba","Type":"ContainerDied","Data":"c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c"} Mar 09 10:06:35 crc kubenswrapper[4692]: I0309 10:06:35.413561 4692 generic.go:334] "Generic (PLEG): container finished" podID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerID="1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109" exitCode=0 Mar 09 10:06:35 crc kubenswrapper[4692]: I0309 10:06:35.413617 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ck7" event={"ID":"1902263a-5295-4bd6-99fa-c486f77b18ba","Type":"ContainerDied","Data":"1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109"} Mar 09 10:06:36 crc kubenswrapper[4692]: I0309 10:06:36.426786 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ck7" event={"ID":"1902263a-5295-4bd6-99fa-c486f77b18ba","Type":"ContainerStarted","Data":"3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5"} Mar 09 10:06:36 crc kubenswrapper[4692]: I0309 10:06:36.448192 4692 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h4ck7" podStartSLOduration=3.005488765 podStartE2EDuration="4.448123968s" podCreationTimestamp="2026-03-09 10:06:32 +0000 UTC" firstStartedPulling="2026-03-09 10:06:34.40140437 +0000 UTC m=+2795.226139941" lastFinishedPulling="2026-03-09 10:06:35.844039563 +0000 UTC m=+2796.668775144" observedRunningTime="2026-03-09 10:06:36.447772638 +0000 UTC m=+2797.272508219" watchObservedRunningTime="2026-03-09 10:06:36.448123968 +0000 UTC m=+2797.272859599" Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.149983 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-t2242/must-gather-xtjbn"] Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.150783 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-t2242/must-gather-xtjbn" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerName="copy" containerID="cri-o://ba26b4e27cf6d0cac52062c9020cae468498a112d80f52770fbc456cefe98679" gracePeriod=2 Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.158595 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-t2242/must-gather-xtjbn"] Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.444707 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t2242_must-gather-xtjbn_6b0f0035-f0b8-48e7-86e8-68187ac429f0/copy/0.log" Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.449149 4692 generic.go:334] "Generic (PLEG): container finished" podID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerID="ba26b4e27cf6d0cac52062c9020cae468498a112d80f52770fbc456cefe98679" exitCode=143 Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.609492 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t2242_must-gather-xtjbn_6b0f0035-f0b8-48e7-86e8-68187ac429f0/copy/0.log" Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.610211 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.700425 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwsld\" (UniqueName: \"kubernetes.io/projected/6b0f0035-f0b8-48e7-86e8-68187ac429f0-kube-api-access-dwsld\") pod \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.700614 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b0f0035-f0b8-48e7-86e8-68187ac429f0-must-gather-output\") pod \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\" (UID: \"6b0f0035-f0b8-48e7-86e8-68187ac429f0\") " Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.810374 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b0f0035-f0b8-48e7-86e8-68187ac429f0-kube-api-access-dwsld" (OuterVolumeSpecName: "kube-api-access-dwsld") pod "6b0f0035-f0b8-48e7-86e8-68187ac429f0" (UID: "6b0f0035-f0b8-48e7-86e8-68187ac429f0"). InnerVolumeSpecName "kube-api-access-dwsld". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.816635 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b0f0035-f0b8-48e7-86e8-68187ac429f0-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "6b0f0035-f0b8-48e7-86e8-68187ac429f0" (UID: "6b0f0035-f0b8-48e7-86e8-68187ac429f0"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.904986 4692 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/6b0f0035-f0b8-48e7-86e8-68187ac429f0-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:37 crc kubenswrapper[4692]: I0309 10:06:37.905065 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwsld\" (UniqueName: \"kubernetes.io/projected/6b0f0035-f0b8-48e7-86e8-68187ac429f0-kube-api-access-dwsld\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:38 crc kubenswrapper[4692]: I0309 10:06:38.097025 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" path="/var/lib/kubelet/pods/6b0f0035-f0b8-48e7-86e8-68187ac429f0/volumes" Mar 09 10:06:38 crc kubenswrapper[4692]: I0309 10:06:38.482924 4692 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-t2242_must-gather-xtjbn_6b0f0035-f0b8-48e7-86e8-68187ac429f0/copy/0.log" Mar 09 10:06:38 crc kubenswrapper[4692]: I0309 10:06:38.486795 4692 scope.go:117] "RemoveContainer" containerID="ba26b4e27cf6d0cac52062c9020cae468498a112d80f52770fbc456cefe98679" Mar 09 10:06:38 crc kubenswrapper[4692]: I0309 10:06:38.487065 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-t2242/must-gather-xtjbn" Mar 09 10:06:38 crc kubenswrapper[4692]: I0309 10:06:38.570419 4692 scope.go:117] "RemoveContainer" containerID="67a68637f9be92f1d4a07c1d7c1b38576c485e2ae27e5c1a99c00c23d628d107" Mar 09 10:06:42 crc kubenswrapper[4692]: I0309 10:06:42.871984 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:42 crc kubenswrapper[4692]: I0309 10:06:42.873321 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:42 crc kubenswrapper[4692]: I0309 10:06:42.916978 4692 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:43 crc kubenswrapper[4692]: I0309 10:06:43.582446 4692 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:43 crc kubenswrapper[4692]: I0309 10:06:43.638604 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ck7"] Mar 09 10:06:45 crc kubenswrapper[4692]: I0309 10:06:45.562214 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h4ck7" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="registry-server" containerID="cri-o://3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5" gracePeriod=2 Mar 09 10:06:45 crc kubenswrapper[4692]: I0309 10:06:45.984831 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.069961 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-catalog-content\") pod \"1902263a-5295-4bd6-99fa-c486f77b18ba\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.070098 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lfb8\" (UniqueName: \"kubernetes.io/projected/1902263a-5295-4bd6-99fa-c486f77b18ba-kube-api-access-5lfb8\") pod \"1902263a-5295-4bd6-99fa-c486f77b18ba\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.070333 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-utilities\") pod \"1902263a-5295-4bd6-99fa-c486f77b18ba\" (UID: \"1902263a-5295-4bd6-99fa-c486f77b18ba\") " Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.071319 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-utilities" (OuterVolumeSpecName: "utilities") pod "1902263a-5295-4bd6-99fa-c486f77b18ba" (UID: "1902263a-5295-4bd6-99fa-c486f77b18ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.077690 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1902263a-5295-4bd6-99fa-c486f77b18ba-kube-api-access-5lfb8" (OuterVolumeSpecName: "kube-api-access-5lfb8") pod "1902263a-5295-4bd6-99fa-c486f77b18ba" (UID: "1902263a-5295-4bd6-99fa-c486f77b18ba"). InnerVolumeSpecName "kube-api-access-5lfb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.100646 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1902263a-5295-4bd6-99fa-c486f77b18ba" (UID: "1902263a-5295-4bd6-99fa-c486f77b18ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.172087 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lfb8\" (UniqueName: \"kubernetes.io/projected/1902263a-5295-4bd6-99fa-c486f77b18ba-kube-api-access-5lfb8\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.172236 4692 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-utilities\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.172421 4692 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1902263a-5295-4bd6-99fa-c486f77b18ba-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.598188 4692 generic.go:334] "Generic (PLEG): container finished" podID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerID="3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5" exitCode=0 Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.598269 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ck7" event={"ID":"1902263a-5295-4bd6-99fa-c486f77b18ba","Type":"ContainerDied","Data":"3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5"} Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.598607 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h4ck7" event={"ID":"1902263a-5295-4bd6-99fa-c486f77b18ba","Type":"ContainerDied","Data":"665286daee24548b9307ca9824dfdb00731fa02ba1b1a181ea7b8201d7f60a69"} Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.598791 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h4ck7" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.599139 4692 scope.go:117] "RemoveContainer" containerID="3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.629925 4692 scope.go:117] "RemoveContainer" containerID="1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.642341 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ck7"] Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.649591 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h4ck7"] Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.676423 4692 scope.go:117] "RemoveContainer" containerID="c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.698633 4692 scope.go:117] "RemoveContainer" containerID="3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5" Mar 09 10:06:46 crc kubenswrapper[4692]: E0309 10:06:46.700030 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5\": container with ID starting with 3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5 not found: ID does not exist" containerID="3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.700070 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5"} err="failed to get container status \"3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5\": rpc error: code = NotFound desc = could not find container \"3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5\": container with ID starting with 3a354e9341e0936054bded6d1b2639cdfee2f995854bbed51a7b2d9c2c28fcf5 not found: ID does not exist" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.700096 4692 scope.go:117] "RemoveContainer" containerID="1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109" Mar 09 10:06:46 crc kubenswrapper[4692]: E0309 10:06:46.700494 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109\": container with ID starting with 1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109 not found: ID does not exist" containerID="1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.700560 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109"} err="failed to get container status \"1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109\": rpc error: code = NotFound desc = could not find container \"1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109\": container with ID starting with 1f8152307f252bd48f4b10637c5c5004871aa70c58091c66feeba2eb9a63a109 not found: ID does not exist" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.700607 4692 scope.go:117] "RemoveContainer" containerID="c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c" Mar 09 10:06:46 crc kubenswrapper[4692]: E0309 10:06:46.701063 4692 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c\": container with ID starting with c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c not found: ID does not exist" containerID="c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c" Mar 09 10:06:46 crc kubenswrapper[4692]: I0309 10:06:46.701100 4692 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c"} err="failed to get container status \"c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c\": rpc error: code = NotFound desc = could not find container \"c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c\": container with ID starting with c210e4ce9ce768dcb77fff8017969632ae54ff992677639cb412659a00172c0c not found: ID does not exist" Mar 09 10:06:48 crc kubenswrapper[4692]: I0309 10:06:48.081614 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" path="/var/lib/kubelet/pods/1902263a-5295-4bd6-99fa-c486f77b18ba/volumes" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.206061 4692 scope.go:117] "RemoveContainer" containerID="e9755eeb3813eec41173bf96202e4a5a090bbaba593bd94d74363956398a3064" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.246704 4692 scope.go:117] "RemoveContainer" containerID="db53439628543717a86a7cf97a9bafae4fecd55f1573c6554af0ea00935d73a1" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.279809 4692 scope.go:117] "RemoveContainer" containerID="ac4efa8079008baaf554c3ede0ef2424c4e673d99d05f101f3e6c486864f1d27" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.311752 4692 scope.go:117] "RemoveContainer" containerID="5ab7049a990d6a34b2fd27408c6fe7c669aba8e49d9d36f8bf91cbbd1569d964" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.347021 4692 scope.go:117] "RemoveContainer" containerID="f369fed400d69274d22d79a1e9711d75aac725496a6acf9315c98aed71d2508c" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.387726 4692 scope.go:117] "RemoveContainer" containerID="1b197625f7ebd47c1a1c9b1d61e82bcda624a5305d0f6315a1ee4595981726aa" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.438917 4692 scope.go:117] "RemoveContainer" containerID="45ddd12681fd658c3d4fa14b32797701a5d87cdb091b32dfb4b5a9e88a9eb338" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.495890 4692 scope.go:117] "RemoveContainer" containerID="419051724af30ea3bdc1e1de61ef54d2dcc223abd644181bed2e505e86edf69f" Mar 09 10:07:09 crc kubenswrapper[4692]: I0309 10:07:09.563533 4692 scope.go:117] "RemoveContainer" containerID="23dd09656a52382798ccf3887d78903320e412c04837135dae56a4e336d51d55" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.144560 4692 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29550848-rw5xg"] Mar 09 10:08:00 crc kubenswrapper[4692]: E0309 10:08:00.145840 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerName="copy" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.145862 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerName="copy" Mar 09 10:08:00 crc kubenswrapper[4692]: E0309 10:08:00.145897 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="extract-content" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.145905 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="extract-content" Mar 09 10:08:00 crc kubenswrapper[4692]: E0309 10:08:00.145921 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="registry-server" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.145930 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="registry-server" Mar 09 10:08:00 crc kubenswrapper[4692]: E0309 10:08:00.145945 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerName="gather" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.145952 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerName="gather" Mar 09 10:08:00 crc kubenswrapper[4692]: E0309 10:08:00.145970 4692 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="extract-utilities" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.145978 4692 state_mem.go:107] "Deleted CPUSet assignment" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="extract-utilities" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.146187 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerName="gather" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.146202 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0f0035-f0b8-48e7-86e8-68187ac429f0" containerName="copy" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.146231 4692 memory_manager.go:354] "RemoveStaleState removing state" podUID="1902263a-5295-4bd6-99fa-c486f77b18ba" containerName="registry-server" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.146849 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550848-rw5xg" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.149527 4692 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-bxv5s" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.150140 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.152544 4692 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.155645 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550848-rw5xg"] Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.238644 4692 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hksl\" (UniqueName: \"kubernetes.io/projected/aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8-kube-api-access-2hksl\") pod \"auto-csr-approver-29550848-rw5xg\" (UID: \"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8\") " pod="openshift-infra/auto-csr-approver-29550848-rw5xg" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.340871 4692 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hksl\" (UniqueName: \"kubernetes.io/projected/aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8-kube-api-access-2hksl\") pod \"auto-csr-approver-29550848-rw5xg\" (UID: \"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8\") " pod="openshift-infra/auto-csr-approver-29550848-rw5xg" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.359820 4692 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hksl\" (UniqueName: \"kubernetes.io/projected/aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8-kube-api-access-2hksl\") pod \"auto-csr-approver-29550848-rw5xg\" (UID: \"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8\") " pod="openshift-infra/auto-csr-approver-29550848-rw5xg" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.465833 4692 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550848-rw5xg" Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.962827 4692 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29550848-rw5xg"] Mar 09 10:08:00 crc kubenswrapper[4692]: I0309 10:08:00.981394 4692 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 09 10:08:01 crc kubenswrapper[4692]: I0309 10:08:01.254429 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550848-rw5xg" event={"ID":"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8","Type":"ContainerStarted","Data":"f853cac89c7f2bc2da5b559f516b242e152ea73429afb354a6b829e54a36211d"} Mar 09 10:08:03 crc kubenswrapper[4692]: I0309 10:08:03.272741 4692 generic.go:334] "Generic (PLEG): container finished" podID="aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8" containerID="00b4927f10a12675ffa04271be71c88d19253acc07519576f4e1c2cd123f386c" exitCode=0 Mar 09 10:08:03 crc kubenswrapper[4692]: I0309 10:08:03.272835 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550848-rw5xg" event={"ID":"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8","Type":"ContainerDied","Data":"00b4927f10a12675ffa04271be71c88d19253acc07519576f4e1c2cd123f386c"} Mar 09 10:08:04 crc kubenswrapper[4692]: I0309 10:08:04.580585 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550848-rw5xg" Mar 09 10:08:04 crc kubenswrapper[4692]: I0309 10:08:04.727634 4692 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hksl\" (UniqueName: \"kubernetes.io/projected/aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8-kube-api-access-2hksl\") pod \"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8\" (UID: \"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8\") " Mar 09 10:08:04 crc kubenswrapper[4692]: I0309 10:08:04.735668 4692 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8-kube-api-access-2hksl" (OuterVolumeSpecName: "kube-api-access-2hksl") pod "aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8" (UID: "aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8"). InnerVolumeSpecName "kube-api-access-2hksl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 09 10:08:04 crc kubenswrapper[4692]: I0309 10:08:04.829744 4692 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hksl\" (UniqueName: \"kubernetes.io/projected/aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8-kube-api-access-2hksl\") on node \"crc\" DevicePath \"\"" Mar 09 10:08:05 crc kubenswrapper[4692]: I0309 10:08:05.292937 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29550848-rw5xg" event={"ID":"aaeb04a3-6e1b-4f5f-b02e-9bc906f9f5d8","Type":"ContainerDied","Data":"f853cac89c7f2bc2da5b559f516b242e152ea73429afb354a6b829e54a36211d"} Mar 09 10:08:05 crc kubenswrapper[4692]: I0309 10:08:05.293001 4692 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f853cac89c7f2bc2da5b559f516b242e152ea73429afb354a6b829e54a36211d" Mar 09 10:08:05 crc kubenswrapper[4692]: I0309 10:08:05.293038 4692 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29550848-rw5xg" Mar 09 10:08:05 crc kubenswrapper[4692]: I0309 10:08:05.651456 4692 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29550842-jtclt"] Mar 09 10:08:05 crc kubenswrapper[4692]: I0309 10:08:05.660736 4692 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29550842-jtclt"] Mar 09 10:08:06 crc kubenswrapper[4692]: I0309 10:08:06.083107 4692 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="536ee033-b615-4af9-b26e-81da38bdc053" path="/var/lib/kubelet/pods/536ee033-b615-4af9-b26e-81da38bdc053/volumes" Mar 09 10:08:09 crc kubenswrapper[4692]: I0309 10:08:09.796496 4692 scope.go:117] "RemoveContainer" containerID="3b246ebbc47316d3bec9ff2fd16ed0332af23549f429cd76dede64fcef70c878" Mar 09 10:08:09 crc kubenswrapper[4692]: I0309 10:08:09.860270 4692 scope.go:117] "RemoveContainer" containerID="effa3c186bf8f4ecb1a079539c061fca884e09581bc30f871f87090bacf83ea2" Mar 09 10:08:09 crc kubenswrapper[4692]: I0309 10:08:09.907822 4692 scope.go:117] "RemoveContainer" containerID="dfeaf99e5565cdce727a23b58d8e19eecb34722c4a27e39bdacca69aa9b733b2" Mar 09 10:08:09 crc kubenswrapper[4692]: I0309 10:08:09.943788 4692 scope.go:117] "RemoveContainer" containerID="770a4b8b6e13f687f65aa53102eee5c4e7ea702e73a0c2d74d586c71ca32055f" Mar 09 10:08:09 crc kubenswrapper[4692]: I0309 10:08:09.989316 4692 scope.go:117] "RemoveContainer" containerID="827267f37ac23a596a19c539b1f3aa69786724c2f77958aa18d7d2e6f3f232b8" Mar 09 10:08:10 crc kubenswrapper[4692]: I0309 10:08:10.015689 4692 scope.go:117] "RemoveContainer" containerID="a310a9938246877ac6786c3964684c3c472aaa879eb1e1d2e494f4cd396ed6b1" Mar 09 10:08:10 crc kubenswrapper[4692]: I0309 10:08:10.042617 4692 scope.go:117] "RemoveContainer" containerID="6b5465a8edc1c16c3d97c7b025ede82fafc800af58001ecd555d91cf9c5786d6" Mar 09 10:08:10 crc kubenswrapper[4692]: I0309 10:08:10.064953 4692 scope.go:117] "RemoveContainer" containerID="84d8c48309a4d506c7678cd9f43585f04c95518a564ac198a40f544bdeef0325" Mar 09 10:08:17 crc kubenswrapper[4692]: I0309 10:08:17.624394 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 10:08:17 crc kubenswrapper[4692]: I0309 10:08:17.625242 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 10:08:47 crc kubenswrapper[4692]: I0309 10:08:47.624613 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 10:08:47 crc kubenswrapper[4692]: I0309 10:08:47.625535 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.624468 4692 patch_prober.go:28] interesting pod/machine-config-daemon-ndpxq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.626935 4692 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.627088 4692 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.628114 4692 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb"} pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.628483 4692 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" containerName="machine-config-daemon" containerID="cri-o://56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb" gracePeriod=600 Mar 09 10:09:17 crc kubenswrapper[4692]: E0309 10:09:17.762382 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.926957 4692 generic.go:334] "Generic (PLEG): container finished" podID="cb18850a-c45f-438b-9854-5f8ced802c58" containerID="56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb" exitCode=0 Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.927036 4692 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" event={"ID":"cb18850a-c45f-438b-9854-5f8ced802c58","Type":"ContainerDied","Data":"56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb"} Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.927146 4692 scope.go:117] "RemoveContainer" containerID="0c4aca92fab9686fc64015cb64a831b1e43d53698f0ab8c3789ce9a9e0a536fa" Mar 09 10:09:17 crc kubenswrapper[4692]: I0309 10:09:17.928045 4692 scope.go:117] "RemoveContainer" containerID="56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb" Mar 09 10:09:17 crc kubenswrapper[4692]: E0309 10:09:17.928600 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:09:31 crc kubenswrapper[4692]: I0309 10:09:31.072794 4692 scope.go:117] "RemoveContainer" containerID="56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb" Mar 09 10:09:31 crc kubenswrapper[4692]: E0309 10:09:31.074089 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" Mar 09 10:09:46 crc kubenswrapper[4692]: I0309 10:09:46.071360 4692 scope.go:117] "RemoveContainer" containerID="56d93302c1ee31c719102b69a013a942a10998ae0a83d6a3be24e5f55a2b7dcb" Mar 09 10:09:46 crc kubenswrapper[4692]: E0309 10:09:46.072543 4692 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ndpxq_openshift-machine-config-operator(cb18850a-c45f-438b-9854-5f8ced802c58)\"" pod="openshift-machine-config-operator/machine-config-daemon-ndpxq" podUID="cb18850a-c45f-438b-9854-5f8ced802c58" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515153516176024457 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015153516176017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015153510016016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015153510017015453 5ustar corecore